emacs-elpa-diffs
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[nongnu] elpa/gptel cdb07d0d2b: gptel: Update description and bump versi


From: ELPA Syncer
Subject: [nongnu] elpa/gptel cdb07d0d2b: gptel: Update description and bump version
Date: Wed, 1 May 2024 18:59:43 -0400 (EDT)

branch: elpa/gptel
commit cdb07d0d2bb6a4fe6d6a620aa4eec98f70fe6f78
Author: Karthik Chikmagalur <karthikchikmagalur@gmail.com>
Commit: Karthik Chikmagalur <karthikchikmagalur@gmail.com>

    gptel: Update description and bump version
    
    gptel.el (header): Update description and bump version.
---
 gptel.el | 36 +++++++++++++++++++++++++++++-------
 1 file changed, 29 insertions(+), 7 deletions(-)

diff --git a/gptel.el b/gptel.el
index 58c74b82e2..17a22e384a 100644
--- a/gptel.el
+++ b/gptel.el
@@ -3,7 +3,7 @@
 ;; Copyright (C) 2023  Karthik Chikmagalur
 
 ;; Author: Karthik Chikmagalur
-;; Version: 0.8.5
+;; Version: 0.8.6
 ;; Package-Requires: ((emacs "27.1") (transient "0.4.0") (compat "29.1.4.1"))
 ;; Keywords: convenience
 ;; URL: https://github.com/karthink/gptel
@@ -32,7 +32,7 @@
 ;; gptel supports
 ;;
 ;; - The services ChatGPT, Azure, Gemini, Anthropic AI, Anyscale, Together.ai,
-;;   Perplexity, and Kagi (FastGPT & Summarizer)
+;;   Perplexity, Anyscale, OpenRouter, Groq and Kagi (FastGPT & Summarizer)
 ;; - Local models via Ollama, Llama.cpp, Llamafiles or GPT4All
 ;;
 ;;  Additionally, any LLM service (local or remote) that provides an
@@ -54,10 +54,14 @@
 ;;   key or to a function of no arguments that returns the key.  (It tries to
 ;;   use `auth-source' by default)
 ;;
+;;   ChatGPT is configured out of the box.  For the other sources:
+;;
 ;; - For Azure: define a gptel-backend with `gptel-make-azure', which see.
 ;; - For Gemini: define a gptel-backend with `gptel-make-gemini', which see.
 ;; - For Anthropic (Claude): define a gptel-backend with 
`gptel-make-anthropic',
 ;;   which see
+;; - For Together.ai, Anyscale, Perplexity, Groq and OpenRouter: define a
+;;   gptel-backend with `gptel-make-openai', which see.
 ;; - For Kagi: define a gptel-backend with `gptel-make-kagi', which see.
 ;;
 ;; For local models using Ollama, Llama.cpp or GPT4All:
@@ -65,6 +69,7 @@
 ;; - The model has to be running on an accessible address (or localhost)
 ;; - Define a gptel-backend with `gptel-make-ollama' or `gptel-make-gpt4all',
 ;;   which see.
+;; - Llama.cpp or Llamafiles: Define a gptel-backend with `gptel-make-openai',
 ;;
 ;; Consult the package README for examples and more help with configuring
 ;; backends.
@@ -81,14 +86,14 @@
 ;; - Call `gptel-send' to send the text up to the cursor.  Select a region to
 ;;   send only the region.
 ;;
-;; - You can select previous prompts and responses to
-;;   continue the conversation.
+;; - You can select previous prompts and responses to continue the 
conversation.
 ;;
 ;; - Call `gptel-send' with a prefix argument to access a menu where you can 
set
 ;;   your backend, model and other parameters, or to redirect the
 ;;   prompt/response.
 ;;
 ;; To use this in a dedicated buffer:
+;; 
 ;; - M-x gptel: Start a chat session
 ;; - C-u M-x gptel: Start another session or multiple independent chat sessions
 ;;
@@ -98,11 +103,28 @@
 ;;   model, or choose to redirect the input or output elsewhere (such as to the
 ;;   kill ring).
 ;;
-;; - You can save this buffer to a file.  When opening this file, turning on
-;;   `gptel-mode' will allow resuming the conversation.
+;; - You can save this buffer to a file.  When opening this file, turn on
+;;   `gptel-mode' before editing it to restore the conversation state and
+;;   continue chatting.
+;;
+;; gptel in Org mode:
+;;
+;; gptel offers a few extra conveniences in Org mode.
+;; - You can limit the conversation context to an Org heading with
+;;   `gptel-org-set-topic'.
+;;   
+;; - You can have branching conversations in Org mode, where each hierarchical
+;;   outline path through the document is a separate conversation branch.
+;;   See the variable `gptel-org-branching-context'.
+;;   
+;; - You can declare the gptel model, backend, temperature, system message and
+;;   other parameters as Org properties with the command
+;;   `gptel-org-set-properties'.  gptel queries under the corresponding heading
+;;   will always use these settings, allowing you to create mostly reproducible
+;;   LLM chat notebooks.
 ;;
 ;; Finally, gptel offers a general purpose API for writing LLM ineractions
-;; that suit how you work, see `gptel-request'.
+;; that suit your workflow, see `gptel-request'.
 
 ;;; Code:
 (declare-function markdown-mode "markdown-mode")



reply via email to

[Prev in Thread] Current Thread [Next in Thread]