Adding expert observation to automated validation
We needed a way to measure the gap between automated validation and expert observation
The DI bundle includes 7 specialized agents. Two were selected as observers for this experiment based on their relevance to document quality.
| Dimension | v0 | v3 | Change |
|---|---|---|---|
| Voice Consistency | 8 | 8 | — |
| Clarity | 9 | 8 | ↓1 |
| Readability | 8 | 7 | ↓1 |
| Information Architecture | 8 | 8 | — |
| Content Flow | 7 | 8 | ↑1 |
| Navigation | 7 | 9 | ↑2 |
Scores are from a single document generation run. These illustrate the type of insight DI feedback provides, not statistically validated results across multiple documents.
These insights were identified in one document generation run. Additional documents may surface different patterns.
Results from a single document generation observation. Promising direction, not proven at scale.
Data as of: February 20, 2026
Feature status: Experimental
Research performed:
Gaps: Scores are from a single document generation run, not statistically validated across multiple documents. Results are illustrative, not definitive.
Primary contributor: Brian Krabach (100% of visible commits)
Consider incorporating DI insights into future outline refinement — addressing issues before generation, not just observing after.