Skip to content

Installation

Requirements

  • Python 3.11+
  • An OpenAI-compatible LLM endpoint (OpenAI, Azure, Ollama, vLLM, GitHub Models, llama.cpp, etc.)

From PyPI

pip install executionkit

This pulls zero runtime dependencies — the default backend is stdlib urllib.

With connection pooling (httpx)

For high-throughput workloads (e.g. consensus with many samples, or long react_loop chains), install the optional httpx extra:

pip install "executionkit[httpx]"

The Provider automatically detects httpx at import time and uses an httpx.AsyncClient with connection pooling. With the stdlib backend, every LLM call opens a fresh TCP+TLS connection.

Verify the install

python -c "from executionkit import Provider, consensus, refine_loop, react_loop; print('OK')"

From source (development)

git clone https://github.com/tafreeman/executionkit.git
cd executionkit
python -m venv .venv
source .venv/bin/activate     # Windows: .venv\Scripts\activate
pip install -e ".[dev]"

The [dev] extra adds pytest, ruff, mypy, bandit, and pre-commit. See Contributing for the full dev workflow.

Run the test suite

pytest -m "not integration"                              # unit tests, no API keys
pytest --cov=executionkit --cov-fail-under=80            # full suite with coverage
OPENAI_API_KEY=sk-... pytest -m integration              # live API tests

Next