Files
vikingowl 2f32d4b954 chore: remove Mistral/Ollama legacy references after Gemini migration
Rename mistral.go → llm_enricher.go and mistral_test.go →
llm_enricher_test.go; update all test function names and stale model
strings (mistral-large-latest → gemini-2.5-flash-lite); drop Ollama
block from .env; mark superseded planning specs; update provider
references in planning docs and CLAUDE.md to Google Gemini.
2026-04-25 17:31:58 +02:00

3.3 KiB

discovery-eval

CLI that grades discovery's AI-backed components against labelled fixtures. Two modes:

  • -mode similarity (default) — SimilarityClassifier on pair- labelled fixtures. Reports precision / recall / F1 / accuracy + a confidence calibration table.
  • -mode categoryProviderLLMEnricher's category output on row- labelled fixtures. Reports accuracy + a per-label confusion matrix.

File-based cache keeps reruns free. Each mode has its own cache key shape, so switching modes doesn't churn entries.

Run it

Similarity (default)

export AI_API_KEY=...
export AI_MODEL_COMPLEX=gemini-2.5-flash-lite

go run ./backend/cmd/discovery-eval \
  -mode similarity \
  -fixture backend/cmd/discovery-eval/fixtures/similarity.json \
  -cache   .eval-cache.json \
  -threshold 0.8 \
  -report  eval-report.json

Exit code is 1 when F1 < threshold (0 = gating disabled).

Category

export AI_API_KEY=...
export AI_MODEL_COMPLEX=gemini-2.5-flash-lite

go run ./backend/cmd/discovery-eval \
  -mode category \
  -fixture backend/cmd/discovery-eval/fixtures/category.json \
  -cache   .cat-eval-cache.json \
  -threshold 0.7 \
  -report  cat-eval-report.json

Category mode scrapes each row's quellen URLs live (first run only; cache covers subsequent runs) and asks the LLM enricher to produce a category. Normalised comparison: casing + German umlauts + the -märkte/-markt plural drift are all treated as equal. Exit code is 1 when accuracy < threshold.

Extending the fixture

fixtures/similarity.json is hand-curated. Add pairs that exercise patterns the classifier has gotten wrong, or edge cases the prompt documentation mentions but we haven't tested. Prefer real crawler output (anonymised if needed) over invented pairs — the model's failure modes on real data are the ones that matter.

Keep pairs deterministic (don't include confidence bands or other stochastic signals) and ensure each pair's note field explains the edge case. When you change a pair's labels, purge the cache so the classifier re-answers.

Cache

-cache .eval-cache.json stores verdicts by sha256(name_normalized_a | stadt_a | year_a | name_normalized_b | stadt_b | year_b) + "|" + model.

Delete the file to force a full re-run. Changing the model string invalidates every entry automatically.

Atomic writes (temp file + rename) — a crashed run won't corrupt the cache.

Interpreting the output

  • Precision is about false merges: how often does the classifier say "same" when it isn't? Low precision means auto-merge will fuse distinct markets — hard to recover from.
  • Recall is about missed merges: how often does it say "different" when it should have said "same"? Low recall means operators review duplicate rows manually — annoying but safe.
  • F1 balances both. The default gate threshold isn't set here — pick a number after an initial baseline run.
  • Calibration tells you whether "90% confident" verdicts are actually correct 90% of the time. An under-calibrated model is worse than a less-accurate but well-calibrated one for downstream auto-decisions.

Out of scope for now

  • Enrichment eval (category/opening_hours/description). Scoring fuzzy text outputs needs its own design; tracked for a future MR.
  • CI wiring. Once we have a baseline F1 the harness can run in GitLab CI with a fixed threshold.