The evaluation pipeline
Every time you runvectorlint doc.md, three things happen in sequence.
1. Rule resolution
VectorLint reads your.vectorlint.ini to determine which rule packs apply to the file. It loads those rule files, prepends the contents of VECTORLINT.md (if present) to each rule’s system prompt, and assembles the evaluation context.
If no .vectorlint.ini exists, VectorLint detects VECTORLINT.md and creates a synthetic “Style Guide Compliance” rule automatically. This is the zero-config path — the fastest way to get from nothing to a working evaluation without writing any rule pack files.
2. LLM evaluation
VectorLint sends your content to the configured LLM provider with each rule’s prompt. Depending on the rule type, the model returns either:- A list of specific violations (check rules) — for countable errors like grammar mistakes or banned terms
- A scored rubric (judge rules) — for quality dimensions like tone, clarity, or technical accuracy rated on a 1–4 scale
Concurrency limit set in .vectorlint.ini.
3. Filtering
Raw model output contains noise — potential violations that don’t hold up under scrutiny. VectorLint reduces false positives through a two-phase filtering process:- Candidate generation — the model returns all potential violations, each tagged with required fields: rule support, exact evidence, context support, plausible non-violation, and fix quality.
- Deterministic surfacing — VectorLint applies a strict filter and only surfaces violations that pass all required gates.
CONFIDENCE_THRESHOLD (default: 0.75). Lower values surface more findings with higher recall; higher values surface fewer findings with higher precision. See Configuring LLM providers for details.
Scoring
VectorLint uses two scoring methods depending on the rule type. Density-based scoring (check rules) calculates error density, violations per 100 words. VectorLint weights a single error in a short document more heavily than the same error in a long one. This normalizes quality assessment across documents of any length. Rubric-based scoring (judge rules) normalizes the LLM’s 1–4 rating to a 1–10 scale and computes a weighted average across all criteria. Criteria weights reflect your real-world priorities — technical accuracy might carry a weight of 40 while SEO carries a weight of 10. See Quality scoring for the full scoring reference.Two ways to define quality
VectorLint gives you two complementary tools for expressing your content standards:VECTORLINT.md— plain-language style instructions that apply globally to every evaluation. The fastest path to useful output: no rule files, no configuration syntax, just plain English instructions that the LLM uses as context for every check.- Rule pack files — structured LLM prompts for specific, measurable checks with weighted scoring criteria. Use these when you need reproducible, auditable results on a particular dimension of quality.
VECTORLINT.md sets the baseline context, and rule pack files enforce specific criteria on top of it.
See Defining your style rules for how to create both.
Next steps
- Quickstart — run your first evaluation in minutes
- Defining your style rules — create VECTORLINT.md and write custom rules
- Configuring a project — map files to rule packs