Stuart Henderson <s...@spacehopper.org> writes: > On 2025/01/30 08:15, Dave Voutila wrote: >> >> FWIW we should be able to include Vulkan support as its in ports. I've >> played with llama.cpp locally with it, but I don't have a GPU that's >> worth a damn top see if it's an improvement over pure CPU-based >> inferencing. > > Makes sense, though I think it would be better to commit without and > add that later. > >> Also should this be arm64 and amd64 specific? I'm not a ports person so >> not sure :) > > Do you mean for llama.cpp at all, or just the vulkan support? > (If it's "at all", afaik the original intention was that - like > whisper.cpp - it would run without anything special).
I think some of its cpu-based inferencing relies on specific cpu extensions, like AVX. Not sure it's truly cross-platform. I may be wrong. > > > On 2025/01/30 05:50, Chris Cappuccio wrote: >> Stuart Henderson [s...@spacehopper.org] wrote: >> > >> > I don't understand why it's in emulators. Perhaps misc would make sense? >> > >> >> I guess either misc or even a new category, like ml. Torch wuold come next, >> and there are plenty of other pieces that really don't fit in any other >> category except misc. > > I'd be happy with misc. If we end up with dozens of related ports then > maybe a new category makes sense but misc seems to fit and is not over-full.