Metadata-Version: 2.4
Name: limin
Version: 0.9.1
Summary: A Python library for interacting with OpenAI-compatible LLM APIs.
Author-email: Titanom Solutions <infrastructure@titanom.com>
Requires-Python: >=3.10
Description-Content-Type: text/markdown
Requires-Dist: openai==1.69.0
Requires-Dist: tqdm==4.67.1

# limin

The `limin` library is a Python library for interacting with OpenAI-compatible LLM APIs.

The central philosophy of `limin` is that interactions with LLMs are a **leaky abstraction**.
This is why `limin` is designed to be as transparent as possible.
You can inspect every object and trace every operation.
The library doesn't try to hide the low-level details behind fancy abstractions - you are in control.

Features:

✅ Parallel generation of text completions.

✅ Timestamping of text completions and measuring generation duration.

✅ Pretty printing of conversations.

✅ Improved type safety and type inference.

✅ Working with log probabilities of tokens (including pretty printing).

✅ Full structured completion support.

✅ Tool call support.

✅ Agentic workflows.

## Installation

Install the library using pip:

```bash
python -m pip install limin
```

## A Simple Example

After you've installed the library, you can use it by importing the `limin` module and calling the functions you need.
You will need to provide the `OPENAI_API_KEY` environment variable.

Now, you can create a simple script that generates a text completion for a user prompt:

```python
from limin import generate_text_completion


async def main():
    completion = await generate_text_completion("What is the capital of France?")
    print(completion.content)


if __name__ == "__main__":
    import asyncio
    asyncio.run(main())
```

This will print something like:

```
The capital of France is Paris.
```

You can find the full example in the [`examples/single_completion.py`](examples/single_completion.py) file.

### Side Note: Passing the API Key

The `limin` library gives you three ways to provide the API key.

The simplest one is to simply set the `OPENAI_API_KEY` environment variable by running `export OPENAI_API_KEY=$YOUR_API_KEY`.

You can also create a `.env` file in the root directory of your project and add the following line:

```
OPENAI_API_KEY=$YOUR_API_KEY
```

Note that you will need to load the `.env` file in your project using a library like `python-dotenv`:

```python
import dotenv

dotenv.load_dotenv()
```

You can also pass the API key to the various functions by passing the `api_key` parameter.
For example:

```python
completion = await generate_text_completion(
    "What is the capital of France?",
    api_key="your_api_key",
)
```

## Generating Text Completions

### Generating a Single Text Completion

You can generate a single text completion for a user prompt by calling the `generate_text_completion` function:

```python
from limin import generate_text_completion

completion = await generate_text_completion("What is the capital of France?")
print(completion.content)
```

You can generate a single text completion for a conversation by calling the `generate_text_completion_for_conversation` function:

```python
from limin import generate_text_completion_for_conversation

conversation = Conversation(
    messages=[
        Message(role="system", content="You are a helpful assistant."),
        Message(role="user", content="What is the capital of France?"),
        Message(role="assistant", content="The capital of France is Paris."),
        Message(role="user", content="What is the capital of Germany?"),
    ]
)
completion = await generate_text_completion_for_conversation(conversation)
print(completion.content)
```

### Generating Multiple Text Completions

You can generate multiple text completions for a list of user prompts by calling the `generate_text_completions` function:

```python
from limin import generate_text_completions

completions = await generate_text_completions([
    "What is the capital of France?",
    "What is the capital of Germany?",
])

for completion in completions:
    print(completion.content)
```

It's important to note that the `generate_text_completions` function will parallelize the generation of the text completions.
The number of parallel completions is controlled by the `n_parallel` parameter (which defaults to 5).

For example, if you want to generate 4 text completions with 2 parallel completions, you can do the following:

```python
completions = await generate_text_completions([
    "What is the capital of France?",
    "What is the capital of Germany?",
    "What is the capital of Italy?",
    "What is the capital of Spain?",
], n_parallel=2)

for completion in completions:
    print(completion.content)
```

You can also generate multiple text completions for a list of conversations by calling the `generate_text_completions_for_conversations` function:

```python
from limin import generate_text_completions_for_conversations

first_conversation = Conversation(
    messages=[
        Message(role="system", content="You are a helpful assistant."),
        Message(role="user", content="What is the capital of France?"),
    ]
)

second_conversation = Conversation(
    messages=[
        Message(role="system", content="You are a helpful assistant."),
        Message(role="user", content="What is the capital of Germany?"),
    ]
)

completions = await generate_text_completions_for_conversations([
    first_conversation,
    second_conversation,
], n_parallel=2)

for completion in completions:
    print(completion.content)
```

Note that both the `generate_text_completions` and `generate_text_completions_for_conversations` functions will show a progress bar if the `show_progress` parameter is set to `True` (which it is by default).
You can suppress this by setting the `show_progress` parameter to `False`.

You can find the full example in the [`examples/multiple_completions.py`](examples/multiple_completions.py) file.

## Generating Structured Completions

You can generate structured completions by calling the equivalent `structured_completion` functions.

For example, you can generate a structured completion for a single user prompt by calling the `generate_structured_completion` function:

```python
from limin import generate_structured_completion

# Note that you need to create a pydantic model containing the expected completion
class CapitalModel(BaseModel):
    capital: str

completion = await generate_structured_completion(
    "What is the capital of France?",
    response_model=CapitalModel,
)
print(completion.content.capital)
```

You can similarly call the `generate_structured_completion_for_conversation`, `generate_structured_completions_for_conversations`, and `generate_structured_completions` functions.
Structured completions also support extracting log probabilities of tokens.

You can find the full example in the [`examples/structured_completion.py`](examples/structured_completion.py) file.

## Tool Call Support

You can generate tool calls by calling the `generate_tool_call_completion` function:

```python
from pydantic import BaseModel, Field
from limin.tool_call import Tool, generate_tool_call_completion


class GetWeatherParameters(BaseModel):
    location: str = Field(description="City and country e.g. Bogotá, Colombia")


get_weather_tool = Tool(
    name="get_weather",
    description="Get current temperature for provided location in celsius.",
    parameters=GetWeatherParameters,
)

completion = await generate_tool_call_completion(
    "What's the weather like in Paris today?",
    get_weather_tool,
)
print(completion)


if __name__ == "__main__":
    asyncio.run(main())
```

You can find the full example in the [`examples/tool_call.py`](examples/tool_call.py) file.

You can also pass multiple tools to the `generate_tool_call_completion` function and have the model choose the best tool to use based on the user prompt.
You can find an example in the [`examples/multiple_tool_calls.py`](examples/multiple_tool_calls.py) file.

Additionally, the `generate_tool_call_completion_for_conversation` function can be used to generate tool calls for a conversation.

## Extracting Log Probabilities

You can extract the log probabilities of the tokens by accessing the `token_log_probs` attribute of the `TextCompletion` object.
You will need to pass the `log_probs` parameter to the generation function together with the `top_log_probs` parameter to get the most likely tokens:

```python
completion = await generate_text_completion(
    "What is 2+2?",
    log_probs=True,
    top_log_probs=10,
)
print(completion.token_log_probs)
```

This will return a list of `TokenLogProb` objects, which have the following attributes:

- `token`: The token.
- `log_prob`: The log probability of the token.

You can pretty print the log probabilities by calling the `to_pretty_log_probs_string` method of the `TextCompletion` object:

```python
print(completion.to_pretty_log_probs_string(show_probabilities=True))
```

This will return a nicely colored string with the log probabilities of the tokens.

You can also access the full list of log probabilities by accessing the `full_token_log_probs` attribute of the `TextCompletion` object:

```python
print(completion.full_token_log_probs)
```

This will return a list of lists of `TokenLogProb` objects (for each token position the `top_log_probs` number of most likely tokens).

You can find the full example in the [`examples/log_probabilities.py`](examples/log_probabilities.py) file.

## Agentic Workflows

You can use the `Agent` class to create agentic workflows.

You can find an example in the [`examples/agent.py`](examples/agent.py) file and an example notebook in the [`notebooks/agent.ipynb`](notebooks/agent.ipynb) file.

## Important Classes

### The Message Class

The `Message` class is a simple dataclass that represents a message in a conversation.
It has the following attributes:

- `role`: The role of the message (either "system", "user", or "assistant").
- `content`: The content of the message.

### The Conversation Class

The `Conversation` class represents a conversation between a user and an assistant.
It contains the `messages` attribute, which is a list of `Message` objects.

You can add a message to the conversation using the `add_message` method.
This will intelligently check whether the message has the correct role and then add the message to the conversation.

Additionally, the `Conversation` class has a `to_pretty_string` method that returns a pretty string representation of the conversation with colored roles and separators.

### The TextCompletion Class

The generation functions return either a `TextCompletion` object or a list of `TextCompletion` objects.
This has the following attributes:

- `conversation`: The conversation that was used to generate the completion.
- `model`: The model that was used to generate the completion.
- `message`: The message that was generated.
- `start_time`: The start time of the generation.
- `end_time`: The end time of the generation.
- `duration`: The duration of the generation took (in seconds).

The `start_time`, `end_time`, and `duration` attributes allow you to benchmark the performance of the generation.

### The ModelConfiguration Class

You can specify the model configuration by passing a `ModelConfiguration` object to the generation functions.

```python
from limin import ModelConfiguration

model_configuration = ModelConfiguration(
    model="gpt-4o",
    temperature=0.7,
    log_probs=True,
    top_log_probs=10,
)

completion = await generate_text_completion(
    "What is 2+2?",
    model_configuration=model_configuration,
)
print(completion.content)
```

You can find the full example in the [`examples/model_configuration.py`](examples/model_configuration.py) file.
