Closed-Book vs Open-Book
Controlled evaluation variants that test the same legal task under different information conditions to distinguish between parametric knowledge and document synthesis.
When a legal AI fails, the cause matters. Did it lack knowledge, or did it fail to find information? Most benchmarks conflate these. LegalChain isolates them by running the exact same task with and without source documents.
"task": "Compare Citations",
"anchor": "384 U.S. 436",
"authority": "418 U.S. 683",
"context": null
}
The performance delta between modalities reveals the Reasoning Bridge. A model excelling closed-book demonstrates genuine legal understanding. Strong open-book performance reveals document analysis capability—extracting info from text it has never encountered.
"task": "Compare Citations",
"anchor": "384 U.S. 436",
"authority": "418 U.S. 683",
"context": "RESEARCH_PACK_V2"
}