Ollama Buddy - Now On MELPA!

[James Dyer] : Feb 24, 2025 : 228 words
emacs linux 🏷️ emacs 2025
 ___ _ _      n _ n      ___       _   _ _ _
|   | | |__._|o(Y)o|__._| . |_ _ _| |_| | | |
| | | | | .  |     | .  | . | | | . | . |__ |
|___|_|_|__/_|_|_|_|__/_|___|___|___|___|___|

https://github.com/captainflasmr/ollama-buddy

A friendly Emacs interface for interacting with Ollama models. This package provides a convenient way to integrate Ollama’s local LLM capabilities directly into your Emacs workflow with little or no configuration required.

Latest improvements:

  • Chat buffer now more prompt based rather than ad-hoc using C-c C-c to send and C-c C-k to cancel
  • Connection monitor now optional, ollama status visibility now maintained by strategic status checks simplifying setup.
  • Can now change models from chat buffer using C-c C-m
  • Updated intro message with ascii logo
  • Suggested default “C-c o” for ollama-buddy-menu
  • defcustom ollama-buddy-command-definitions now will work in the customization interface.
  • The presets directory on github contains elisp files that can be evaluated to generate a role-based menu.
  • Added to MELPA, install using the following:
(use-package ollama-buddy
  :bind ("C-c o" . ollama-buddy-menu))
  • and to add initial model:
(use-package ollama-buddy
   :bind ("C-c o" . ollama-buddy-menu)
   :custom ollama-buddy-default-model "llama3.2:1b")

Roadmap :

  • DOING Multi-shot prompt to multiple LLMS to choose best answer

  • DOING Customizable Role-Based Menu Preset Generation System

  • DOING Distinguishing commands associated with LLMs using colours

For any more information just have a look at the github README!