If you use Ollama and don’t want to write your own API wrapper every time,
the ollama_call project significantly simplifies the work.
This is a small Python library that allows you to send a request to a local LLM with one function
and immediately receive a response, including in JSON format.
Installation
pip install ollama-call
Why is it needed
- minimal code for working with the model;
- structured JSON response for further processing;
- convenient for rapid prototypes and MVPs;
- supports streaming output if necessary.
Use example
from ollama_call import ollama_call
response = ollama_call(
user_prompt="Hello, how are you?",
format="json",
model="gemma3:12b"
)
print(response)
When it is especially useful
- you write scripts or services on top of Ollama;
- need a predictable response format;
- there is no desire to connect heavy frameworks.
Total
ollama_call is a lightweight and clear wrapper for working with Ollama from Python.
A good choice if simplicity and quick results are important.
GitHub
https://github.com/demensdeum/ollama_call