On 2025/01/29 21:12, Chris Cappuccio wrote: > This is a simple port that works well out-of-the-box. It's kind of an > emulator, > I think. > > It has some optional python scripts that will need numpy and torch. > > I'm not sure the right way to handle the version since it's "b4589" and not > 1.0. > The ports tree thinks that's too incomplete, root/stem differentiation.
Here are diffs to go on top. This sets the package version to 0.0.4589 which is closely enough related I think. What you had will break things because it puts something other than a subpackage name in the place where a subpackage name should go. Also: use shlib versioning, enable tests, disable ccache autodetect (ports handles ccache on it's own and internal support usually messes this up), stop using -march=native (no good for bulk builds). I don't understand why it's in emulators. Perhaps misc would make sense? diff -uNp -r llama.cpp/Makefile /usr/ports/emulators/llama.cpp/Makefile --- llama.cpp/Makefile Wed Jan 29 23:36:30 2025 +++ /usr/ports/emulators/llama.cpp/Makefile Thu Jan 30 09:03:29 2025 @@ -1,5 +1,16 @@ COMMENT = LLM inference system +GH_ACCOUNT = ggerganov +GH_PROJECT = llama.cpp +GH_TAGNAME = b4589 +PKGNAME = llama-cpp-0.0.${GH_TAGNAME:S/b//} + +SHARED_LIBS += ggml-base 0.0 +SHARED_LIBS += ggml-cpu 0.0 +SHARED_LIBS += ggml 0.0 +SHARED_LIBS += llama 0.0 +SHARED_LIBS += llava_shared 0.0 + CATEGORIES = emulators HOMEPAGE = https://github.com/ggerganov/llama.cpp @@ -9,15 +20,8 @@ PERMIT_PACKAGE = Yes WANTLIB += m pthread ${COMPILER_LIBCXX} -GH_ACCOUNT = ggerganov -GH_PROJECT = llama.cpp -GH_TAGNAME = b4589 - -V = 0.0 -FULLPKGNAME = llama-cpp-${GH_TAGNAME}-${V} - MODULES = devel/cmake - -NO_TEST = Yes +CONFIGURE_ARGS = -DGGML_CCACHE=Off \ + -DGGML_NATIVE=Off .include <bsd.port.mk> diff -uNp -r llama.cpp/pkg/PLIST /usr/ports/emulators/llama.cpp/pkg/PLIST --- llama.cpp/pkg/PLIST Wed Jan 29 23:19:25 2025 +++ /usr/ports/emulators/llama.cpp/pkg/PLIST Thu Jan 30 08:47:24 2025 @@ -81,9 +81,9 @@ lib/cmake/ggml/ggml-version.cmake lib/cmake/llama/ lib/cmake/llama/llama-config.cmake lib/cmake/llama/llama-version.cmake -@so lib/libggml-base.so -@so lib/libggml-cpu.so -@so lib/libggml.so -@so lib/libllama.so -@so lib/libllava_shared.so +@lib lib/libggml-base.so.${LIBggml-base_VERSION} +@lib lib/libggml-cpu.so.${LIBggml-cpu_VERSION} +@lib lib/libggml.so.${LIBggml_VERSION} +@lib lib/libllama.so.${LIBllama_VERSION} +@lib lib/libllava_shared.so.${LIBllava_shared_VERSION} lib/pkgconfig/llama.pc