pretty good - only things left to do are library bumps, and register a
dependency on curl. I'll take care of it and commit later this evening.


On 2025/05/14 17:33, Percy Piper wrote:
> Hi.
> 
> Fairly heavy user of this on amd64 -current (Vulkan).
> 
> Many useful improvements including support for newer models (Gemini3, QWEN3
> etc.) and llama-server now exits gracefully from ^C.
> 
> I don't have much clue about ports, so apologies for any errors. No current
> maintainer so sending directly to ports@ - hope that's okay.
> 
> (https://github.com/ggml-org/llama.cpp/pull/13541)
> 
> Percy.
> 
> 
> 
> Index: Makefile
> ===================================================================
> RCS file: /cvs/ports/misc/llama.cpp/Makefile,v
> diff -u -p -r1.7 Makefile
> --- Makefile  18 Feb 2025 00:02:17 -0000      1.7
> +++ Makefile  14 May 2025 16:05:05 -0000
> @@ -8,7 +8,7 @@ COMMENT =             LLM inference system
> 
>  GH_ACCOUNT =         ggerganov
>  GH_PROJECT =         llama.cpp
> -GH_TAGNAME =         b4706
> +GH_TAGNAME =         b5372
>  PKGNAME =            llama-cpp-0.0.${GH_TAGNAME:S/b//}
> 
>  SHARED_LIBS +=               ggml-base 0.0
> Index: distinfo
> ===================================================================
> RCS file: /cvs/ports/misc/llama.cpp/distinfo,v
> diff -u -p -r1.2 distinfo
> --- distinfo  13 Feb 2025 12:21:58 -0000      1.2
> +++ distinfo  14 May 2025 16:05:05 -0000
> @@ -1,2 +1,2 @@
> -SHA256 (llama.cpp-b4706.tar.gz) =
> jpINppeW9Vu/jeqf9gnJPsZ1Hkpkj6YWOHbJSAcPwxc=
> -SIZE (llama.cpp-b4706.tar.gz) = 20705861
> +SHA256 (llama.cpp-b5372.tar.gz) =
> 28q/8fqCc/rtzo8qbUtQ6njW/Zs3mzaLH0ynCWIkuk8=
> +SIZE (llama.cpp-b5372.tar.gz) = 21147804
> Index: patches/patch-common_common_cpp
> ===================================================================
> RCS file: patches/patch-common_common_cpp
> diff -N patches/patch-common_common_cpp
> --- /dev/null 1 Jan 1970 00:00:00 -0000
> +++ patches/patch-common_common_cpp   14 May 2025 16:05:05 -0000
> @@ -0,0 +1,12 @@
> +Index: common/common.cpp
> +--- common/common.cpp.orig
> ++++ common/common.cpp
> +@@ -830,7 +830,7 @@ std::string fs_get_cache_directory() {
> +     if (getenv("LLAMA_CACHE")) {
> +         cache_directory = std::getenv("LLAMA_CACHE");
> +     } else {
> +-#if defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)
> ++#if defined(__linux__) || defined(__FreeBSD__) || defined(_AIX) ||
> defined(__OpenBSD__)
> +         if (std::getenv("XDG_CACHE_HOME")) {
> +             cache_directory = std::getenv("XDG_CACHE_HOME");
> +         } else {
> Index: patches/patch-tools_rpc_rpc-server_cpp
> ===================================================================
> RCS file: patches/patch-tools_rpc_rpc-server_cpp
> diff -N patches/patch-tools_rpc_rpc-server_cpp
> --- /dev/null 1 Jan 1970 00:00:00 -0000
> +++ patches/patch-tools_rpc_rpc-server_cpp    14 May 2025 16:05:05 -0000
> @@ -0,0 +1,12 @@
> +Index: tools/rpc/rpc-server.cpp
> +--- tools/rpc/rpc-server.cpp.orig
> ++++ tools/rpc/rpc-server.cpp
> +@@ -111,7 +111,7 @@ static std::string fs_get_cache_directory() {
> +     if (getenv("LLAMA_CACHE")) {
> +         cache_directory = std::getenv("LLAMA_CACHE");
> +     } else {
> +-#if defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)
> ++#if defined(__linux__) || defined(__FreeBSD__) || defined(_AIX) ||
> defined(__OpenBSD__)
> +         if (std::getenv("XDG_CACHE_HOME")) {
> +             cache_directory = std::getenv("XDG_CACHE_HOME");
> +         } else {
> Index: pkg/PLIST
> ===================================================================
> RCS file: /cvs/ports/misc/llama.cpp/pkg/PLIST,v
> diff -u -p -r1.3 PLIST
> --- pkg/PLIST 13 Feb 2025 12:21:59 -0000      1.3
> +++ pkg/PLIST 14 May 2025 16:05:05 -0000
> @@ -8,28 +8,23 @@ bin/convert_hf_to_gguf.py
>  @bin bin/llama-embedding
>  @bin bin/llama-eval-callback
>  @bin bin/llama-export-lora
> -@bin bin/llama-gbnf-validator
> +@bin bin/llama-finetune
>  @bin bin/llama-gen-docs
>  @bin bin/llama-gguf
>  @bin bin/llama-gguf-hash
>  @bin bin/llama-gguf-split
>  @bin bin/llama-gritlm
>  @bin bin/llama-imatrix
> -@bin bin/llama-infill
> -@bin bin/llama-llava-cli
> -@bin bin/llama-llava-clip-quantize-cli
>  @bin bin/llama-lookahead
>  @bin bin/llama-lookup
>  @bin bin/llama-lookup-create
>  @bin bin/llama-lookup-merge
>  @bin bin/llama-lookup-stats
> -@bin bin/llama-minicpmv-cli
> +@bin bin/llama-mtmd-cli
>  @bin bin/llama-parallel
>  @bin bin/llama-passkey
>  @bin bin/llama-perplexity
>  @bin bin/llama-quantize
> -@bin bin/llama-quantize-stats
> -@bin bin/llama-qwen2vl-cli
>  @bin bin/llama-retrieval
>  @bin bin/llama-run
>  @bin bin/llama-save-load-state
> @@ -45,6 +40,7 @@ include/ggml-alloc.h
>  include/ggml-backend.h
>  include/ggml-blas.h
>  include/ggml-cann.h
> +include/ggml-cpp.h
>  include/ggml-cpu.h
>  include/ggml-cuda.h
>  include/ggml-kompute.h
> @@ -68,5 +64,5 @@ lib/cmake/llama/llama-version.cmake
>  @lib lib/libggml-vulkan.so.${LIBggml-vulkan_VERSION}
>  @lib lib/libggml.so.${LIBggml_VERSION}
>  @lib lib/libllama.so.${LIBllama_VERSION}
> -@lib lib/libllava_shared.so.${LIBllava_shared_VERSION}
> +@so lib/libmtmd_shared.so
>  lib/pkgconfig/llama.pc


Reply via email to