From 42f7e201af301696066a792832e3b47c5597a6fc Mon Sep 17 00:00:00 2001 From: flu0r1ne Date: Sat, 6 May 2023 05:52:00 -0500 Subject: Update readme / bump version --- README.md | 25 ++++++++++++++++--------- pyproject.toml | 2 +- setup.py | 3 ++- src/gpt_chat_cli/version.py | 2 +- 4 files changed, 20 insertions(+), 12 deletions(-) diff --git a/README.md b/README.md index 83e3a0d..653c2dc 100644 --- a/README.md +++ b/README.md @@ -3,7 +3,7 @@ gpt-chat-cli > This is a WIP. Please consider it experimental. -A simple ChatGPT CLI. +A simple yet powerful ChatGPT CLI. ### Installation: @@ -76,12 +76,12 @@ Note that this implementation uses the `powi()` method to compute the squared di ### Usage: ``` -usage: gpt-chat-cli [-h] [-m MODEL] [-t TEMPERATURE] [-f FREQUENCY_PENALTY] [-p PRESENCE_PENALTY] [-k MAX_TOKENS] [-s TOP_P] [-n N_COMPLETIONS] - [--adornments {AutoDetectedOption.ON,AutoDetectedOption.OFF,AutoDetectedOption.AUTO}] [--color {AutoDetectedOption.ON,AutoDetectedOption.OFF,AutoDetectedOption.AUTO}] - message +usage: gcli.py [-h] [-m MODEL] [-t TEMPERATURE] [-f FREQUENCY_PENALTY] [-p PRESENCE_PENALTY] [-k MAX_TOKENS] [-s TOP_P] [-n N_COMPLETIONS] [--adornments {on,off,auto}] [--color {on,off,auto}] [--version] [-l] + [-i] + [message] positional arguments: - message The contents of the message. When used in chat mode, this is the initial message if provided. + message The contents of the message. When in a interactive session, this is the initial prompt provided. options: -h, --help show this help message and exit @@ -100,10 +100,13 @@ options: comprising the top 10% probability mass are considered. -n N_COMPLETIONS, --n-completions N_COMPLETIONS How many chat completion choices to generate for each input message. - --adornments {AutoDetectedOption.ON,AutoDetectedOption.OFF,AutoDetectedOption.AUTO} + --adornments {on,off,auto} Show adornments to indicate the model and response. Can be set to 'on', 'off', or 'auto'. - --color {AutoDetectedOption.ON,AutoDetectedOption.OFF,AutoDetectedOption.AUTO} + --color {on,off,auto} Set color to 'on', 'off', or 'auto'. + --version Print version and exit + -l, --list-models List models and exit + -i, --interactive Start an interactive session ``` ### Features: @@ -112,5 +115,9 @@ options: - [x] Color and adornments - [x] Support for multiple completions - [x] Support for any model which is supported through the chat completions API. [See model endpoint compatibility.](https://platform.openai.com/docs/models/model-endpoint-compatibility) -- [x] Capability to modify parameters including temperature, frequency penalty, presence penalty, top p, and the maximum number of tokens emitted. - +- [x] Capability to modify parameters including temperature, frequency penalty, presence penalty, top p, and the maximum number of tokens emitted +- [x] Interactive sessions +- [x] Code syntax highlighting (when the LLM labels code blocks) +- [x] Interactive mode with emacs-style editing +- [x] Input from standard in, pipes, and heredoc +- [x] List available models diff --git a/pyproject.toml b/pyproject.toml index 462d249..f445e40 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta" [project] name = "gpt-chat-cli" -version = "0.0.1" +version = "0.1.0" authors = [ { name="Flu0r1ne", email="flu0r1ne@flu0r1ne.net" }, ] diff --git a/setup.py b/setup.py index 0576ace..0c78604 100644 --- a/setup.py +++ b/setup.py @@ -2,7 +2,7 @@ import setuptools setuptools.setup( name='gpt-chat-cli', - version='0.0.1', + version='0.1.0', entry_points = { 'console_scripts': ['gpt-chat-cli=gpt_chat_cli.gcli:main'], }, @@ -13,6 +13,7 @@ setuptools.setup( install_requires=[ 'setuptools', 'openai >= 0.27.6', + 'pygments >= 0.15.0' ], python_requires='>=3.7' ) diff --git a/src/gpt_chat_cli/version.py b/src/gpt_chat_cli/version.py index bc784a3..1f6518e 100644 --- a/src/gpt_chat_cli/version.py +++ b/src/gpt_chat_cli/version.py @@ -1 +1 @@ -VERSION = '0.0.2' +VERSION = '0.1.0' -- cgit v1.2.3