User Guide

Practical how-to for daily use, based on the current codebase.

Overview

This app is a local-first assistant. The browser UI is served by your own Flask server and talks to local models through Ollama. Speech features can run in the browser or on the server.

Chats, audio, and analytics stay on your machine by default. Caddy provides the HTTPS gateway so the UI can be used securely from a phone while Ollama stays bound to localhost.

What the app includes

This is the current scope of the product:

Getting Started

Benchmark Workflow

Use benchmark pages as a three-step flow: baseline report, alternative interpretation, then live monitor when a run is active.

  1. Read the primary report: Open /docs/benchmark_guided.html for the canonical benchmark summary and model ranking.
  2. Compare with alternative analysis: Open /docs/benchmark_autonomous_claude.html to cross-check conclusions using a different KPI weighting and chart style.
  3. Track an active run live: Open /docs/benchmark_monitor.html for real-time progress, telemetry, and the latest executed task.

When to use which page:

Access Levels and Limits

These limits keep the app responsive when many people connect at once. They protect the GPU and avoid long queues while still letting everyone explore the app safely.

Guest — a safe, read-only account for public access and demos.

Normal users — full chat access with guardrails to keep the server smooth.

Admins — unrestricted access for owners.

Conversation Basics

Type in the composer and press Send or Enter. Use Shift+Enter for line breaks. The composer grows with your text.

Modes and Models

Use the Fast / Normal / Deep buttons in the header to control how detailed the response should be.

Sessions and History

Speech Input (STT)

Speech Output (TTS)

Appearance and Settings

Admin and Debug Tools

Keyboard Shortcuts

Remote Access Notes

If the app is exposed publicly, access still requires login for all API routes. The docs are public.

Troubleshooting