You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
swarms/docs/swarms/models/openai.md

196 lines
8.5 KiB

# `BaseOpenAI` and `OpenAI` Documentation
## Table of Contents
1. [Overview](#overview)
2. [Class Architecture](#class-architecture)
3. [Purpose](#purpose)
4. [Class Attributes](#class-attributes)
5. [Methods](#methods)
- [Construction](#construction)
- [Configuration](#configuration)
- [Tokenization](#tokenization)
- [Generation](#generation)
- [Asynchronous Generation](#asynchronous-generation)
6. [Usage Examples](#usage-examples)
- [Creating an OpenAI Object](#creating-an-openai-object)
- [Generating Text](#generating-text)
- [Advanced Configuration](#advanced-configuration)
---
## 1. Overview <a name="overview"></a>
The `BaseOpenAI` and `OpenAI` classes are part of the LangChain library, designed to interact with OpenAI's large language models (LLMs). These classes provide a seamless interface for utilizing OpenAI's API to generate natural language text.
## 2. Class Architecture <a name="class-architecture"></a>
Both `BaseOpenAI` and `OpenAI` classes inherit from `BaseLLM`, demonstrating an inheritance-based architecture. This architecture allows for easy extensibility and customization while adhering to the principles of object-oriented programming.
## 3. Purpose <a name="purpose"></a>
The purpose of these classes is to simplify the interaction with OpenAI's LLMs. They encapsulate API calls, handle tokenization, and provide a high-level interface for generating text. By instantiating an object of the `OpenAI` class, developers can quickly leverage the power of OpenAI's models to generate text for various applications, such as chatbots, content generation, and more.
## 4. Class Attributes <a name="class-attributes"></a>
Here are the key attributes and their descriptions for the `BaseOpenAI` and `OpenAI` classes:
| Attribute | Description |
|---------------------------|-------------|
| `lc_secrets` | A dictionary of secrets required for LangChain, including the OpenAI API key. |
| `lc_attributes` | A dictionary of attributes relevant to LangChain. |
| `is_lc_serializable()` | A method indicating if the class is serializable for LangChain. |
| `model_name` | The name of the language model to use. |
| `temperature` | The sampling temperature for text generation. |
| `max_tokens` | The maximum number of tokens to generate in a completion. |
| `top_p` | The total probability mass of tokens to consider at each step. |
| `frequency_penalty` | Penalizes repeated tokens according to frequency. |
| `presence_penalty` | Penalizes repeated tokens. |
| `n` | How many completions to generate for each prompt. |
| `best_of` | Generates `best_of` completions server-side and returns the "best." |
| `model_kwargs` | Holds any model parameters valid for `create` calls not explicitly specified. |
| `openai_api_key` | The OpenAI API key used for authentication. |
| `openai_api_base` | The base URL for the OpenAI API. |
| `openai_organization` | The OpenAI organization name, if applicable. |
| `openai_proxy` | An explicit proxy URL for OpenAI requests. |
| `batch_size` | The batch size to use when passing multiple documents for generation. |
| `request_timeout` | The timeout for requests to the OpenAI completion API. |
| `logit_bias` | Adjustment to the probability of specific tokens being generated. |
| `max_retries` | The maximum number of retries to make when generating. |
| `streaming` | Whether to stream the results or not. |
| `allowed_special` | A set of special tokens that are allowed. |
| `disallowed_special` | A collection of special tokens that are not allowed. |
| `tiktoken_model_name` | The model name to pass to `tiktoken` for token counting. |
## 5. Methods <a name="methods"></a>
### 5.1 Construction <a name="construction"></a>
#### 5.1.1 `__new__(cls, **data: Any) -> Union[OpenAIChat, BaseOpenAI]`
- Description: Initializes the OpenAI object.
- Arguments:
- `cls` (class): The class instance.
- `data` (dict): Additional data for initialization.
- Returns:
- Union[OpenAIChat, BaseOpenAI]: An instance of the OpenAI class.
### 5.2 Configuration <a name="configuration"></a>
#### 5.2.1 `build_extra(cls, values: Dict[str, Any]) -> Dict[str, Any]`
- Description: Builds extra kwargs from additional params passed in.
- Arguments:
- `cls` (class): The class instance.
- `values` (dict): Values and parameters to build extra kwargs.
- Returns:
- Dict[str, Any]: A dictionary of built extra kwargs.
#### 5.2.2 `validate_environment(cls, values: Dict) -> Dict`
- Description: Validates that the API key and python package exist in the environment.
- Arguments:
- `values` (dict): The class values and parameters.
- Returns:
- Dict: A dictionary of validated values.
### 5.3 Tokenization <a name="tokenization"></a>
#### 5.3.1 `get_sub_prompts(self, params: Dict[str, Any], prompts: List[str], stop: Optional[List[str]] = None) -> List[List[str]]`
- Description: Gets sub-prompts for LLM call.
- Arguments:
- `params` (dict): Parameters for LLM call.
- `prompts` (list): List of prompts.
- `stop` (list, optional): List of stop words.
- Returns:
- List[List[str]]: List of sub-prompts.
#### 5.3.2 `get_token_ids(self, text: str) -> List[int]`
- Description: Gets token IDs using the `tiktoken` package.
- Arguments:
- `text` (str): The text for which to calculate token IDs.
- Returns:
- List[int]: A list of token IDs.
#### 5.3.3 `modelname_to_contextsize(modelname: str) -> int`
- Description: Calculates the maximum number of tokens possible to generate for a model.
- Arguments:
- `modelname` (str): The model name to determine the context size for.
- Returns:
- int: The maximum context size.
#### 5.3.4 `max_tokens_for_prompt(self, prompt: str) -> int`
- Description: Calculates the maximum number of tokens possible to generate for a prompt.
- Arguments:
- `prompt` (str): The prompt for which to
determine the maximum token limit.
- Returns:
- int: The maximum token limit.
### 5.4 Generation <a name="generation"></a>
#### 5.4.1 `generate(self, text: Union[str, List[str]], **kwargs) -> Union[str, List[str]]`
- Description: Generates text using the OpenAI API.
- Arguments:
- `text` (str or list): The input text or list of inputs.
- `**kwargs` (dict): Additional parameters for the generation process.
- Returns:
- Union[str, List[str]]: The generated text or list of generated texts.
### 5.5 Asynchronous Generation <a name="asynchronous-generation"></a>
#### 5.5.1 `generate_async(self, text: Union[str, List[str]], **kwargs) -> Union[str, List[str]]`
- Description: Generates text asynchronously using the OpenAI API.
- Arguments:
- `text` (str or list): The input text or list of inputs.
- `**kwargs` (dict): Additional parameters for the asynchronous generation process.
- Returns:
- Union[str, List[str]]: The generated text or list of generated texts.
## 6. Usage Examples <a name="usage-examples"></a>
### 6.1 Creating an OpenAI Object <a name="creating-an-openai-object"></a>
```python
# Import the OpenAI class
from swarms.models import OpenAI
# Set your OpenAI API key
api_key = "YOUR_API_KEY"
# Create an OpenAI object
openai = OpenAI(api_key)
```
### 6.2 Generating Text <a name="generating-text"></a>
```python
# Generate text from a single prompt
prompt = "Translate the following English text to French: 'Hello, how are you?'"
generated_text = openai.generate(prompt, max_tokens=50)
# Generate text from multiple prompts
prompts = ["Translate this: 'Good morning' to Spanish.", "Summarize the following article:", article_text]
generated_texts = openai.generate(prompts, max_tokens=100)
# Generate text asynchronously
async_prompt = "Translate 'Thank you' into German."
async_result = openai.generate_async(async_prompt, max_tokens=30)
# Access the result of an asynchronous generation
async_result_text = async_result.get()
```
### 6.3 Advanced Configuration <a name="advanced-configuration"></a>
```python
# Configure generation with advanced options
custom_options = {
"temperature": 0.7,
"max_tokens": 100,
"top_p": 0.9,
"frequency_penalty": 0.2,
"presence_penalty": 0.4
}
generated_text = openai.generate(prompt, **custom_options)
```
This documentation provides a comprehensive understanding of the `BaseOpenAI` and `OpenAI` classes, their attributes, methods, and usage examples. Developers can utilize these classes to interact with OpenAI's language models efficiently, enabling various natural language generation tasks.