Added
- Code Bench v1 — Multi-model consensus evaluator for code quality, sibling to Doc Bench under Testing & QA → Evaluators
- Project types: Spring Boot, FastAPI, React, Generic
- Languages: Python, TypeScript, JavaScript, Java, Other
- Per-file, classification-aware rubrics keyed by
(language, project_type, classification) - 5-step wizard: Scan → Select → Classify → Configure → Results
- Tri-state file tree, per-file dimension + model breakdown
- Mongo-backed rubric authoring with in-app wizard editor
- Markdown / PDF / CSV export
- SSE streaming evaluation with cancel support
- Markview DMN/DRL viewers — Native rendering for DMN decision tables and Drools DRL files
- Code Insights user scoping — Runs now scoped by user email and machine ID for multi-user installations
Changed
- Evaluators grouping — Doc Bench and Code Bench unified under a single Evaluators section
- Code Bench results UX — Refined per-file results presentation, distinct navigation icon
- LLM-only recommendations — Code Bench recommendations now strictly LLM-provided (no rule-based fallback)
- Model registry — Tightened to latest model options across providers
- Progressive markdown rendering — Improved rendering performance for large markdown documents
Fixed
- License activation reliability with v2 safeguards
- Code Bench evaluation progress now visible after page refresh (server-side fallback when SSE hook is cold)
- Code Bench evaluation hook: cancel correctly resolves stuck evaluating state
- Code Bench list page: runs sorted by recency, Space-key activation works for keyboard users
- Code Insights spec tree fixes
Security
- Vulnerable dependency upgrades across the stack
- Tightened release vulnerability gates in CI