branch: externals/llm commit 3a2d6962f37c9b1fd5882206214304667bd481b6 Author: Andrew Hyatt <ahy...@gmail.com> Commit: Andrew Hyatt <ahy...@gmail.com>
Whitespace fixes --- llm-claude.el | 4 ++-- llm-openai.el | 38 +++++++++++++++++++------------------- llm-provider-utils.el | 10 +++++----- llm-tester.el | 2 +- llm-vertex.el | 6 +++--- llm.el | 4 ++-- 6 files changed, 32 insertions(+), 32 deletions(-) diff --git a/llm-claude.el b/llm-claude.el index 7ad49fb410..a14154438f 100644 --- a/llm-claude.el +++ b/llm-claude.el @@ -102,7 +102,7 @@ (cl-defmethod llm-provider-streaming-media-handler ((_ llm-claude) msg-receiver _ err-receiver) (cons 'text/event-stream - (plz-event-source:text/event-stream + (plz-event-source:text/event-stream :events `((message_start . ignore) (content_block_start . ignore) (ping . ignore) @@ -114,7 +114,7 @@ . ,(lambda (event) (let* ((data (plz-event-source-event-data event)) - (json (json-parse-string data :object-type 'alist)) + (json (json-parse-string data :object-type 'alist)) (delta (assoc-default 'delta json)) (type (assoc-default 'type delta))) (when (equal type "text_delta") diff --git a/llm-openai.el b/llm-openai.el index ba9b007b39..af035c4b43 100644 --- a/llm-openai.el +++ b/llm-openai.el @@ -209,14 +209,14 @@ RESPONSE can be nil if the response is complete." (cl-defmethod llm-provider-streaming-media-handler ((_ llm-openai) msg-receiver fc-receiver _) (cons 'text/event-stream - (plz-event-source:text/event-stream + (plz-event-source:text/event-stream :events `((message . - ,(lambda (event) - (let ((data (plz-event-source-event-data event))) - (unless (equal data "[DONE]") - (when-let ((response (llm-openai--get-partial-chat-response - (json-read-from-string data)))) + ,(lambda (event) + (let ((data (plz-event-source-event-data event))) + (unless (equal data "[DONE]") + (when-let ((response (llm-openai--get-partial-chat-response + (json-read-from-string data)))) (funcall (if (stringp response) msg-receiver fc-receiver) response)))))))))) (cl-defmethod llm-provider-collect-streaming-function-data ((_ llm-openai) data) @@ -224,19 +224,19 @@ RESPONSE can be nil if the response is complete." (dotimes (i (length (car data))) (setf (aref cvec i) (make-llm-provider-utils-function-call))) (cl-loop for part in data do - (cl-loop for call in (append part nil) do - (let* ((index (assoc-default 'index call)) - (id (assoc-default 'id call)) - (function (assoc-default 'function call)) - (name (assoc-default 'name function)) - (arguments (assoc-default 'arguments function))) - (when id - (setf (llm-provider-utils-function-call-id (aref cvec index)) id)) - (when name - (setf (llm-provider-utils-function-call-name (aref cvec index)) name)) - (setf (llm-provider-utils-function-call-args (aref cvec index)) - (concat (llm-provider-utils-function-call-args (aref cvec index)) - arguments))))) + (cl-loop for call in (append part nil) do + (let* ((index (assoc-default 'index call)) + (id (assoc-default 'id call)) + (function (assoc-default 'function call)) + (name (assoc-default 'name function)) + (arguments (assoc-default 'arguments function))) + (when id + (setf (llm-provider-utils-function-call-id (aref cvec index)) id)) + (when name + (setf (llm-provider-utils-function-call-name (aref cvec index)) name)) + (setf (llm-provider-utils-function-call-args (aref cvec index)) + (concat (llm-provider-utils-function-call-args (aref cvec index)) + arguments))))) (cl-loop for call in (append cvec nil) do (setf (llm-provider-utils-function-call-args call) (json-read-from-string (llm-provider-utils-function-call-args call))) diff --git a/llm-provider-utils.el b/llm-provider-utils.el index 5b752f50c2..644662840e 100644 --- a/llm-provider-utils.el +++ b/llm-provider-utils.el @@ -198,7 +198,7 @@ return a list of `llm-chat-function-call' structs.") data (or (llm-provider-embedding-extract-error provider data) - "Unknown error"))))))) + "Unknown error"))))))) (cl-defmethod llm-chat ((provider llm-standard-chat-provider) prompt) (llm-provider-request-prelude provider) @@ -254,12 +254,12 @@ return a list of `llm-chat-function-call' structs.") :media-type (llm-provider-streaming-media-handler provider (lambda (s) - (when (> (length s) 0) + (when (> (length s) 0) (setq current-text (concat current-text s)) (when partial-callback - (llm-provider-utils-callback-in-buffer - buf partial-callback current-text)))) + (llm-provider-utils-callback-in-buffer + buf partial-callback current-text)))) (lambda (fc-new) (push fc-new fc)) (lambda (err) (llm-provider-utils-callback-in-buffer @@ -273,7 +273,7 @@ return a list of `llm-chat-function-call' structs.") provider prompt current-text (llm-provider-collect-streaming-function-data - provider (nreverse fc))))) + provider (nreverse fc))))) :on-error (lambda (_ data) (llm-provider-utils-callback-in-buffer buf error-callback 'error diff --git a/llm-tester.el b/llm-tester.el index 31c6cec1a6..4607492c2e 100644 --- a/llm-tester.el +++ b/llm-tester.el @@ -119,7 +119,7 @@ (llm-chat-streaming provider (llm-make-chat-prompt - "Write a medium length poem in iambic pentameter about the pleasures of using Emacs. The poem should make snide references to vi." + "Write a medium length poem in iambic pentameter about the pleasures of using Emacs. The poem should make snide references to vi." :temperature 0.5) (lambda (text) (unless (eq buf (current-buffer)) diff --git a/llm-vertex.el b/llm-vertex.el index 4a4c3f6be8..a69c6e2d48 100644 --- a/llm-vertex.el +++ b/llm-vertex.el @@ -247,8 +247,8 @@ nothing to add, in which case it is nil." :handler (lambda (element) (if-let ((response (llm-provider-chat-extract-result provider element))) - (funcall msg-receiver response) - (when-let ((fc (llm-provider-extract-function-calls provider element))) + (funcall msg-receiver response) + (when-let ((fc (llm-provider-extract-function-calls provider element))) (funcall fc-receiver fc))))))) (cl-defmethod llm-provider-collect-streaming-function-data ((_ llm-google) data) @@ -294,7 +294,7 @@ If STREAMING is non-nil, use the URL for the streaming API." (cl-defmethod llm-count-tokens ((provider llm-google) string) (llm-provider-request-prelude provider) - (let ((response (llm-request-sync + (let ((response (llm-request-sync (llm-google-count-tokens-url provider) :headers (llm-provider-headers provider) :data (llm-vertex--to-count-token-request diff --git a/llm.el b/llm.el index 5053c41c64..dbc853a7fe 100644 --- a/llm.el +++ b/llm.el @@ -79,7 +79,7 @@ ROLE can a symbol, of either `user', `assistant', or `function'. FUNCTION-CALL-RESULTS is a struct of type `llm-chat-prompt-function-call-results', which is only populated if `role' is `function'. It stores the results of just one -function call." +function call." role content function-call-result) (cl-defstruct llm-chat-prompt-function-call-result @@ -379,7 +379,7 @@ be passed to `llm-cancel-request'." (cl-defmethod llm-chat-streaming :around (provider prompt partial-callback response-callback error-callback) "Log the input to llm-chat-async." (llm--log 'api-send :provider provider :prompt prompt) - ;; We need to wrap the callbacks before we set llm-log to nil. + ;; We need to wrap the callbacks before we set llm-log to nil. (let* ((new-partial-callback (lambda (response) (when partial-callback (llm--log 'api-receive-partial :provider provider :msg response)