fieri

Crates.iofieri
lib.rsfieri
version0.7.0
sourcesrc
created_at2022-12-27 09:38:36.921187
updated_at2023-05-02 22:35:19.475709
descriptionUnofficial Rust client for OpenAI.
homepagehttps://docs.rs/fieri
repositoryhttps://github.com/lbkolev/fieri
max_upload_size
id746130
size1,159,571
lachezar kolev (lbkolev)

documentation

https://docs.rs/fieri

README

fieri

Overview

Unofficial Rust client for OpenAI.

fieri provides an asynchronous Rust interface for interaction with OpenAI, allowing you to easily use OpenAI's state-of-the-art machine learning models in your Rust projects.

Prerequisites

Before you can use the Rust Client for OpenAI, you'll need to sign up for an API key at the OpenAI Developer Portal. Once you've signed up, you'll be able to find your API key in the API Keys section of the developer portal.

Installation

Run cargo add fieri in your terminal to add the latest version of the client.

Basic Usage

Generate text based on a prompt

use fieri::{
    completion::{create, CompletionParamBuilder},
    Client, Error,
};
use std::env;

#[tokio::main]
async fn main() -> Result<(), Error> {
    let client = Client::new();

    let param = CompletionParamBuilder::new("ada")
        .prompt("Generate a plot for an absurd interstellar parody.")
        .max_tokens(500)
        .temperature(0.9)
        .top_p(1.0)
        .frequency_penalty(0.0)
        .presence_penalty(0.0)
        .build()?;

    let resp = create(&client, &param).await?;
    println!("Generated text: {:#?}", resp);

    Ok(())
}

Generate and stream back text based on a prompt

use fieri::{
    completion::{create_with_stream, Completion, CompletionParamBuilder},
    Client, Error,
};
use std::env;

#[tokio::main]
async fn main() -> Result<(), Error> {
    let client = Client::new();

    let param = CompletionParamBuilder::new("ada")
        .prompt("unnecessarily lo")
        .temperature(0.5)
        .build()?;

    let mut resp = create_with_stream(&client, &param).await?;

    while let Some(chunk) = resp.chunk().await? {
        if chunk.to_vec() == b"data: [DONE]\n\n" {
            break;
        }

        let v: Completion = serde_json::from_slice(&chunk[5..])?;
        v.choices.iter().for_each(|c| println!("{:?}", c.text));
    }

    Ok(())
}

Generate an image based on a prompt and save it locally.

use fieri::{
    image::{ImageSize, GenerateImageParamBuilder, generate},
    Client, Error,
};
use std::env;

#[tokio::main]
async fn main() -> Result<(), Error> {
    let client = Client::new();

    let param = GenerateImageParamBuilder::new("A bunch of cats dancing tango on top of the highest mountain on Mars.")
        .size(ImageSize::S1024x1024)
        .n(1)
        .build()?;

    generate(&client, &param)
        .await?
        .save("/tmp/")
        .await?;

    Ok(())
}

By default, the api key and organization are implicitly loaded from environment variables OPENAI_API_KEY & OPENAI_ORGANIZATION. it's possible to configure/overwrite them per client. e.g: using for example:

use fieri::Client

let client = Client::new().api_key("<key>");
let client_with_org = Client::new().organization("<organization>");

More examples can be found in the docs.

Limitations

Note that the Rust Client for OpenAI is provided as-is, and is not officially supported by OpenAI. While we will do our best to keep the library up-to-date and bug-free, we cannot guarantee that it will always work as expected.

Additionally, the API has usage limits that may affect your ability to use the models. You can view your current usage and limits in the Usage section of your account.

License

fieri is provided under the MIT license. See LICENSE.

Commit count: 56

cargo fmt