branch: elpa/gptel
commit b2a2c66f2ae6f56057bedfde4d9f367dab69bb5e
Author: Karthik Chikmagalur <karthikchikmaga...@gmail.com>
Commit: Karthik Chikmagalur <karthikchikmaga...@gmail.com>

    gptel: Add support for gpt-4o-mini
    
    * gptel.el (gptel--openai, gptel-model): Add support for
    gpt-4o-mini.
---
 gptel.el | 8 +++++---
 1 file changed, 5 insertions(+), 3 deletions(-)

diff --git a/gptel.el b/gptel.el
index 02b94e5377..e642c00cf6 100644
--- a/gptel.el
+++ b/gptel.el
@@ -412,6 +412,7 @@ by the LLM provider's API.
 The current options for ChatGPT are
 - \"gpt-3.5-turbo\"
 - \"gpt-3.5-turbo-16k\"
+- \"gpt-4o-mini\"
 - \"gpt-4\"
 - \"gpt-4o\"
 - \"gpt-4-turbo\"
@@ -424,6 +425,7 @@ To set the model for a chat session interactively call
   :safe #'always
   :type '(choice
           (string :tag "Specify model name")
+          (const :tag "GPT 4 omni mini" "gpt-4o-mini")
           (const :tag "GPT 3.5 turbo" "gpt-3.5-turbo")
           (const :tag "GPT 3.5 turbo 16k" "gpt-3.5-turbo-16k")
           (const :tag "GPT 4" "gpt-4")
@@ -451,9 +453,9 @@ To set the temperature for a chat session interactively call
    "ChatGPT"
    :key 'gptel-api-key
    :stream t
-   :models '("gpt-3.5-turbo" "gpt-3.5-turbo-16k" "gpt-4" "gpt-4o"
-             "gpt-4-turbo" "gpt-4-turbo-preview" "gpt-4-32k"
-             "gpt-4-1106-preview" "gpt-4-0125-preview")))
+   :models '("gpt-3.5-turbo" "gpt-3.5-turbo-16k" "gpt-4o-mini"
+             "gpt-4" "gpt-4o" "gpt-4-turbo" "gpt-4-turbo-preview"
+             "gpt-4-32k" "gpt-4-1106-preview" "gpt-4-0125-preview")))
 
 (defcustom gptel-backend gptel--openai
   "LLM backend to use.

Reply via email to