branch: elpa/gptel commit b2a2c66f2ae6f56057bedfde4d9f367dab69bb5e Author: Karthik Chikmagalur <karthikchikmaga...@gmail.com> Commit: Karthik Chikmagalur <karthikchikmaga...@gmail.com>
gptel: Add support for gpt-4o-mini * gptel.el (gptel--openai, gptel-model): Add support for gpt-4o-mini. --- gptel.el | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/gptel.el b/gptel.el index 02b94e5377..e642c00cf6 100644 --- a/gptel.el +++ b/gptel.el @@ -412,6 +412,7 @@ by the LLM provider's API. The current options for ChatGPT are - \"gpt-3.5-turbo\" - \"gpt-3.5-turbo-16k\" +- \"gpt-4o-mini\" - \"gpt-4\" - \"gpt-4o\" - \"gpt-4-turbo\" @@ -424,6 +425,7 @@ To set the model for a chat session interactively call :safe #'always :type '(choice (string :tag "Specify model name") + (const :tag "GPT 4 omni mini" "gpt-4o-mini") (const :tag "GPT 3.5 turbo" "gpt-3.5-turbo") (const :tag "GPT 3.5 turbo 16k" "gpt-3.5-turbo-16k") (const :tag "GPT 4" "gpt-4") @@ -451,9 +453,9 @@ To set the temperature for a chat session interactively call "ChatGPT" :key 'gptel-api-key :stream t - :models '("gpt-3.5-turbo" "gpt-3.5-turbo-16k" "gpt-4" "gpt-4o" - "gpt-4-turbo" "gpt-4-turbo-preview" "gpt-4-32k" - "gpt-4-1106-preview" "gpt-4-0125-preview"))) + :models '("gpt-3.5-turbo" "gpt-3.5-turbo-16k" "gpt-4o-mini" + "gpt-4" "gpt-4o" "gpt-4-turbo" "gpt-4-turbo-preview" + "gpt-4-32k" "gpt-4-1106-preview" "gpt-4-0125-preview"))) (defcustom gptel-backend gptel--openai "LLM backend to use.