Go to file
Paul Harrison fc07f3132f Bump to v1.0 2023-09-29 14:28:19 +01:00
src/llm_chat Set default history directory to ~/.llm_chat 2023-09-27 15:33:22 +01:00
tests Set default history directory to ~/.llm_chat 2023-09-27 15:33:22 +01:00
.gitignore Store conversation history on session end 2023-08-24 18:22:15 +01:00
LICENSE Initial commit 2023-08-13 19:53:05 +01:00
Makefile Add linting checks and code formatting 2023-08-16 10:10:06 +01:00
README.md Add installation and usage instructions 2023-09-28 13:14:26 +01:00
llm-chat.png Add installation and usage instructions 2023-09-28 13:14:26 +01:00
poetry.lock Name conversation 2023-09-14 18:02:45 +01:00
pyproject.toml Bump to v1.0 2023-09-29 14:28:19 +01:00

README.md

LLM Chat

This project is currently a CLI interface for OpenAI's GPT model API. The long term goal is to be much more general and interface with any LLM, whether that be a cloud service or a model running locally.

Usage

Installation

The package is not currently published so you will need to build it yourself. To do so you will need Poetry v1.5 or later. Once you have that installed, run the following commands:

git clone https://code.harrison.sh/paul/llm-chat.git
cd llm-chat
poetry build

This will create a wheel file in the dist directory. You can then install it with pip:

pip install dist/llm_chat-0.6.1-py3-none-any.whl

Note, a much better way to install this package system wide is to use pipx. This will install the package in an isolated environment and make the llm command available on your path.

Configuration

Your OpenAI API key must be set in the OPENAI_API_KEY environment variable. The following additional environment variables are supported:

Variable Description
OPENAI_MODEL The model to use. Defaults to gpt-3.5-turbo.
OPENAI_TEMPERATURE The temperature to use when generating text. Defaults to 0.7.
OPENAI_HISTORY_DIR The directory to store chat history in. Defaults to ~/.llm_chat.

The settings can also be configured via command line arguments. Run llm chat --help for more information.

Usage

To start a chat session, run the following command:

llm chat 

This will start a chat session with the default settings looking something like the image below:

Chat session

The CLI accepts multi-line input, so to send the input to the model press Esc + Enter. After each message a running total cost will be displayed. To exit the chat session, send the message /q.

There are several additional options available:

  • --context allows you to provide one or more text files as context for your chat, which is sent as an initial system message at the start of the conversation. You can provide the argument multiple times to add multiple files.
  • --name gives your chat session a name. This is used to name the history file.

See llm chat --help for a full description of all options and their usage.

Finally, you can load a previous chat session using the llm load command. For example:

llm load /path/to/session.json

See llm load --help for more information.