branch: elpa/gptel commit 4c5f7e6953c3e4dac36b4c9cf4b4a2c854febfad Author: Karthik Chikmagalur <karthikchikmaga...@gmail.com> Commit: Karthik Chikmagalur <karthikchikmaga...@gmail.com>
gptel: Bump version to v0.9.8 * gptel.el: Update version and package description. --- gptel.el | 23 ++++++++++++++--------- 1 file changed, 14 insertions(+), 9 deletions(-) diff --git a/gptel.el b/gptel.el index d3e77ce1c3..e8f5e1406a 100644 --- a/gptel.el +++ b/gptel.el @@ -3,7 +3,7 @@ ;; Copyright (C) 2023 Karthik Chikmagalur ;; Author: Karthik Chikmagalur <karthik.chikmaga...@gmail.com> -;; Version: 0.9.7 +;; Version: 0.9.8 ;; Package-Requires: ((emacs "27.1") (transient "0.7.4") (compat "29.1.4.1")) ;; Keywords: convenience, tools ;; URL: https://github.com/karthink/gptel @@ -36,7 +36,7 @@ ;; ;; - The services ChatGPT, Azure, Gemini, Anthropic AI, Anyscale, Together.ai, ;; Perplexity, Anyscale, OpenRouter, Groq, PrivateGPT, DeepSeek, Cerebras, -;; Github Models, xAI and Kagi (FastGPT & Summarizer). +;; Github Models, Novita AI, xAI and Kagi (FastGPT & Summarizer). ;; - Local models via Ollama, Llama.cpp, Llamafiles or GPT4All ;; ;; Additionally, any LLM service (local or remote) that provides an @@ -51,6 +51,7 @@ ;; - Supports conversations and multiple independent sessions. ;; - Supports tool-use to equip LLMs with agentic capabilities. ;; - Supports multi-modal models (send images, documents). +;; - Supports "reasoning" content in LLM responses. ;; - Save chats as regular Markdown/Org/Text files and resume them later. ;; - You can go back and edit your previous prompts or LLM responses when ;; continuing a conversation. These will be fed back to the model. @@ -70,9 +71,11 @@ ;; - For Gemini: define a gptel-backend with `gptel-make-gemini', which see. ;; - For Anthropic (Claude): define a gptel-backend with `gptel-make-anthropic', ;; which see. -;; - For Together.ai, Anyscale, Perplexity, Groq, OpenRouter, DeepSeek, Cerebras or +;; - For Together.ai, Anyscale, Groq, OpenRouter, DeepSeek, Cerebras or ;; Github Models: define a gptel-backend with `gptel-make-openai', which see. ;; - For PrivateGPT: define a backend with `gptel-make-privategpt', which see. +;; - For Perplexity: define a backend with `gptel-make-perplexity', which see. +;; - For Deepseek: define a backend with `gptel-make-deepseek', which see. ;; - For Kagi: define a gptel-backend with `gptel-make-kagi', which see. ;; ;; For local models using Ollama, Llama.cpp or GPT4All: @@ -125,11 +128,12 @@ ;; Include more context with requests: ;; ;; If you want to provide the LLM with more context, you can add arbitrary -;; regions, buffers or files to the query with `gptel-add'. To add text or -;; media files, call `gptel-add' in Dired or use the dedicated `gptel-add-file'. +;; regions, buffers, files or directories to the query with `gptel-add'. To add +;; text or media files, call `gptel-add' in Dired or use the dedicated +;; `gptel-add-file'. ;; -;; You can also add context from gptel's menu instead (gptel-send with a prefix -;; arg), as well as examine or modify context. +;; You can also add context from gptel's menu instead (`gptel-send' with a +;; prefix arg), as well as examine or modify context. ;; ;; When context is available, gptel will include it with each LLM query. ;; @@ -156,8 +160,9 @@ ;; will always use these settings, allowing you to create mostly reproducible ;; LLM chat notebooks. ;; -;; Finally, gptel offers a general purpose API for writing LLM ineractions -;; that suit your workflow, see `gptel-request'. +;; Finally, gptel offers a general purpose API for writing LLM ineractions that +;; suit your workflow. See `gptel-request', and `gptel-fsm' for more advanced +;; usage. ;;; Code: (declare-function markdown-mode "markdown-mode")