Crates.io | gptee |
lib.rs | gptee |
version | 0.3.10 |
source | src |
created_at | 2023-01-29 23:23:27.914972 |
updated_at | 2023-09-02 03:12:13.788287 |
description | output from a language model using standard input as the prompt |
homepage | https://github.com/zurawiki/gptee |
repository | https://github.com/zurawiki/gptee |
max_upload_size | |
id | 771276 |
size | 1,586,761 |
gptee
Output from a language model using standard input as the prompt
Now supporting GPT3.5 chat completions!
cargo
(recommended).cargo install --locked gptee
gptee
was designed for use within shell scripts and other programs, and also works in interactive shells.
Simple example
echo Tell me a joke | gptee
Why did the chicken cross the road?
To get to the other side!
Compose shell commands like you would in a script
echo Tell me a joke | gptee | say
You can compose command and execute them in a script. Proceed with caution before running arbitrary shell scripts
echo Give me just a macOS zsh command to get the free space on my hard drive \
| gptee -s "Prefix each line of output with a pound sign if it not meant to be executed" \
# pipe this to `sh` to have it execute
Try with a custom model. By default gptee
uses gpt-3.5-turbo
echo Tell me a joke | gptee -m text-davinci-003
Using a chat completion model (like gpt-3.5-turbo
), you can then inject a system message with -s
or --system-message
. For davinci and other non-chat models, the output is prefixed to the prompt.
echo "Tell me I'm pretty" | gptee -s "You only speak French"
See the --help
/ -h
flag for more features.
If you encounter any bugs or have any suggestions for improvements, please open an issue on the repository.
This project is licensed under the MIT License.
output from a language model using standard input as the prompt
Usage: gptee [OPTIONS] [FILE]...
Arguments:
[FILE]... File(s) to print / concatenate. Use a dash ('-') or no argument at all to read from standard input
Options:
-m, --model <MODEL>
ID of the model to use. You can use the [List models](https://beta.openai.com/docs/api-reference/models/list) API to see all of your available models, or see our [Model overview](https://beta.openai.com/docs/models/overview) for descriptions of them [default: gpt-3.5-turbo]
-l, --max-tokens <MAX_TOKENS>
The maximum number of [tokens](/tokenizer) to generate in the completion
--stop <STOP>
Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence
-t, --temperature <TEMPERATURE>
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic [default: 0.7]
--top-p <TOP_P>
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered [default: 1]
-s, --system-message <SYSTEM_MESSAGE>
For chat completions, you can specify a system message to be sent to the model. This message will be sent to the model before the user's message. This is useful for providing context to the model, or for providing a prompt to the model. See https://platform.openai.com/docs/guides/chat for more details
-h, --help
Print help (see more with '--help')
-V, --version
Print version