On 2/27/26 16:44, Koen Kooi wrote:
CAUTION: This email comes from a non Wind River email account!
Do not click links or open attachments unless you recognize the sender and know 
the content is safe.

Op 25 feb 2026, om 06:22 heeft Hongxu Jia <[email protected]> het 
volgende geschreven:

On 1/31/26 00:56, Richard Purdie wrote:
CAUTION: This email comes from a non Wind River email account!
Do not click links or open attachments unless you recognize the sender and know 
the content is safe.

Hi Hongxu,

On Fri, 2026-01-30 at 13:20 +0800, hongxu via lists.openembedded.org wrote:

Would you please approve to add this layer to
https://git.yoctoproject.org/, I will maintain
meta-ollama for recipe uprev and bug fix

Thanks for sharing this, it looks really interesting and I like the
idea of it a lot.

Also, thanks for volunteering to maintain it, that does help alleviate
various concerns.

Approval for this rests with the Yocto Project TSC so I have asked them
about it. We should have a decision after our next meeting.

Hi Richard,
Ping, what is the status of the approval from Yocto Project TSC?
About the suggestion to rename the layer to meta-ai or meta-llm and collect 
other LLM applications, such as llama.cpp.
I am open to it, but I strongly insist to maintain meta-ollama as a standalone 
layer, further more, if we support llama.cpp later, we should add it as 
meta-llama-cpp
Then we could customize and isolate each LLM application in a layer, including:
- Add model recipe for LLM application, the format of LLM differs for each 
application (ollama vs llama.cpp)
- Boot up LLM application for Yocto image, add systemd service file

- Customize GPU support (CUDA) for LLM application

If Yocto decides to use meta-ai or meta-llm, I suggest to add meta-ollama and 
meta-llama-cpp as the subset of it, such as meta-xx in meta-openembedded
What are the technical arguments for needing to put those into their own 
layers? As I've stated before, I'd like to have a single layer that has the 
most used AI/ML things and I can help maintaining that on company time. 
Splitting that into multiple layers would increase the friction for both 
consumers and developers of those layers.

As above mentioned, the format of LLM differs for each application, the model of ollama from ollama regristry https://registry.ollama.ai/, the modle of llama.cpp is from huggingface https://huggingface.co/

if we add a model recipe, we have to split the application name if we place them in one layer


About boot up management, ollama has a single server to listen on specific port for all local models,

but llama.cpp start a server to listen one port for one model, it does not support one port for multiple models, and we have software llama-swap https://github.com/mostlygeek/llama-swap to manage multiple models for llama.cpp


Currently ollama and llama.cpp are standalone LLM application and no direct relation with each other, splitting layers for LLM application will help use focus on each LLM application,

It is helpful to classify the maintain scope, based on LLM application.


//Hongxu



regards,

Koen


-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#2292): 
https://lists.openembedded.org/g/openembedded-architecture/message/2292
Mute This Topic: https://lists.openembedded.org/mt/117540395/21656
Group Owner: [email protected]
Unsubscribe: https://lists.openembedded.org/g/openembedded-architecture/unsub 
[[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to