Skip to main content

Quickstart

This guide helps you get llm.port running quickly and send your first successful request.

Prerequisites

  • Docker Engine 24+ with Compose V2
  • 8 GB RAM minimum (16 GB recommended)
  • Python 3.12+ (if using pip-based CLI install)

1) Install the CLI

pip install llmport-cli

You can also use the standalone binary if you prefer not to install Python tooling.

2) Validate your host

llmport doctor

3) Start the platform

llmport up

4) Complete initial setup

Open the admin console and complete the first-run setup:

  • add a provider
  • choose a model alias
  • create or paste an API token

Once done, your apps can call llm.port through one stable API endpoint.

5) Send a test request

curl http://localhost:4000/v1/chat/completions \
-H "Authorization: Bearer <token>" \
-H "Content-Type: application/json" \
-d '{
"model": "<model-alias>",
"messages": [{"role": "user", "content": "Hello"}]
}'

What to do next

  • Configure providers and model aliases
  • Enable modules you need (for example PII or RAG)
  • Review Security Overview

Screenshots

Providers setup

Model catalog

This documentation is generated with AI assistance and may contain inaccuracies. Please validate critical details before production use.