Quick Start
Get LLMOps Eval running locally in minutes.
Prerequisites
| Requirement | Version |
|---|---|
| Java | 21+ |
| Python | 3.11+ |
| Node.js | 18+ |
| PostgreSQL | 16 |
| Docker | Optional |
Option 1: Docker Compose (Recommended)
The fastest way to get started.
git clone https://github.com/ashwithpoojary98/llm-eval-ops.git
cd llm-eval-ops
docker-compose up -d
Access:
- Frontend: http://localhost:3000
- Spring Boot API: http://localhost:8080
- FastAPI Docs: http://localhost:8000/docs
Option 2: Manual Setup
1. Clone the Repository
git clone https://github.com/ashwithpoojary98/llm-eval-ops.git
cd llm-eval-ops
2. Database Setup
# Using Docker
docker run -d \
--name postgres-llmops \
-e POSTGRES_DB=llmevalplatform \
-e POSTGRES_USER=postgres \
-e POSTGRES_PASSWORD=postgres \
-p 5432:5432 \
postgres:16
3. Spring Boot API
# Set environment variables
export SPRING_DATASOURCE_URL=jdbc:postgresql://localhost:5432/llmevalplatform
export SPRING_DATASOURCE_USERNAME=postgres
export SPRING_DATASOURCE_PASSWORD=postgres
export JWT_SECRET_KEY=your-secure-secret-key-at-least-32-chars
export LLMOPS_ADMIN_EMAIL=admin@example.com
export LLMOPS_ADMIN_PASSWORD=ChangeMe123!
export LLMOPS_ORG_NAME="My Organization"
export LLMOPS_ALLOWED_DOMAINS=example.com
export CORS_ALLOWED_ORIGINS=http://localhost:3000
# Run
./mvnw spring-boot:run
API starts on http://localhost:8080
4. FastAPI Evaluation Engine
cd evaluation-engine
python -m venv venv
source venv/bin/activate # Windows: venv\Scripts\activate
pip install -r requirements.txt
cp .env.example .env
# Edit .env with your database URL
python run.py
Evaluation engine starts on http://localhost:8000
5. Next.js Frontend
cd frontend
npm install
# Configure environment
echo "NEXT_PUBLIC_API_URL=http://localhost:8080/api" > .env.local
npm run dev
Frontend starts on http://localhost:3000
First Evaluation
- Open http://localhost:3000
- Log in with your admin credentials
- Create a Project
- Upload a Dataset (CSV or JSON)
- Configure an LLM Endpoint (OpenAI, Anthropic, etc.)
- Select Metrics (BLEU, Faithfulness, etc.)
- Run Evaluation and view results