jeremy ardley wrote:

> Ask ChatGPT4 . Explain what you are trying to do and get it
> to give you a suitable context and prompt

I don't know what to ask exactly, maybe I can ask ChatGPT4 ...

> localdocs contains text you trust that can be used in
> responses in preference to something synthesised from
> the internet.

I tried to google but it seems everything is local with the
llamafile so I don't know, also if it was included it would
have been in the man page or appeared as a simple plugin when
you googled it.

Anyway, the context is big enough to play around with for
now then.

Or the context size is maybe the same? Just when looking for
an answer, the AI first looks for the localdocs, then
the internet? Or you can throw in stuff there, and this gets
crunched by the AI? But then, I don't have that much stuff, so
I would have to get it from the internet anyway.

Wait, let's first find out how big a file you can use! In my
command, we see that -c or --ctx-size "Set the size of the
prompt context" (llamafile(1)).

If the unit is tokens, that means I know have not 512 but 2048
of them. There are a lot of token related stuff in the log
even when it is disabled, seemingly.

--prompt-cache-all will maybe store some data on the file or
when you exit the application. Then that sweat stuff will be
brought just right in again.

taskset -c 0-2       \
./llm                \
  --cli              \
  --log-disable      \
  --prompt-cache-all \
  --silent-prompt    \
  -c 2048            \
  -ngl 9999          \
  -p "$(cat $src)" > $dst

-- 
underground experts united
https://dataswamp.org/~incal

Reply via email to