LegalBenchPro is an EMNLP-targeted research benchmark for evaluating large language models on open-ended legal reasoning. The project asks whether models that perform well on scalable public legal-exam tasks also transfer to de-identified, practice-oriented case analysis.
The benchmark separates two evaluation settings:
- public legal-exam tasks with reference answers;
- de-identified Chinese civil judgment prompts that require stance-aware, statute-grounded legal analysis.
This repository is a public research snapshot. It exposes the dataset schema, scoring protocol, audit workflow, manuscript materials, compact content previews, and summary metadata while keeping the complete workbook private until licensing, privacy, and redistribution review are complete.
- Scope: Chinese institutional and legal text, with both scalable public-exam prompts and de-identified civil-judgment reasoning tasks.
- Evaluation design: comparable task construction, model-configuration metadata, scoring regimes, and staged human-validation plans.
- Reproducibility: Python sample extraction, machine-readable metadata, tests, data-card documentation, and an explicit workflow audit trail.
- Release status: public research preview; full data release is held back pending privacy, licensing, and validation review.
| Component | Current count | Evaluation design |
|---|---|---|
| Chinese real-case split | 76 issue-stance prompts | Citation-aware rubric with human validation in progress |
| Source judgments | 15 de-identified civil judgments | Paired support/opposition issue prompts |
| Public-exam split | 868 instances | Reference-answer consistency scoring |
| Model configurations | 22 | Standard, reasoning-enabled, and step-by-step prompting modes |
| Main multimodel response cells | 20,768 LLM-generated responses | 944 task instances x 22 model configurations |
| Human validation pilots | 10 real-case rows; 80 public-exam rows | Staged for reviewer calibration and agreement analysis |
The public snapshot includes 10 English-language preview rows from the Chinese real-case split, 20 preview rows from the public-exam split, model-configuration metadata, and compact source/domain distribution tables. Preview CSV cells are capped at 420 characters. The repository does not include the full prompt matrix, full reference answers, full model outputs, row-level full indexes, or human review sheets.
LegalBenchPro is designed around a gap in current legal LLM evaluation: public legal benchmarks are scalable and convenient, but legal practice often requires working from long facts, contested interpretations, jurisdiction-specific authorities, and defensible argument structure. This project contributes:
- a two-part benchmark that separates public-exam evaluation from real-case legal analysis;
- a curated Chinese civil judgment split with paired issue-stance prompts;
- a multimodel evaluation matrix spanning 22 model configurations and 20,768 LLM-generated response cells;
- a scoring protocol that distinguishes answer matching from citation-aware legal reasoning;
- a reproducible public workflow for sample extraction, metadata generation, and manuscript tracking.
For a quick review of the project, start with:
paper/introduction_revised.texfor the current manuscript introduction;docs/DATA_CARD.mdfor scope, counts, intended uses, and release constraints;docs/ANNOTATION_PROTOCOL.mdfor human-validation and scoring design;docs/AI_WORKFLOW.mdfor auditability and AI-assistance safeguards;data/README.mdfor a compact public data preview;data/sample/legalbenchpro_cn_judgments_sample.csvfor real-case content excerpts;data/sample/legalbenchpro_public_exam_sample.csvfor public-exam content excerpts;data/metadata/source_distribution.csvanddata/metadata/model_configurations.csvfor concise metadata;scripts/extract_public_sample.pyfor the reproducible sample-export workflow.
paper/
LegalBenchPro_intro_draft.pdf # Overleaf PDF snapshot of the current draft
introduction_revised.tex # Dataset-aligned introduction ready for Overleaf
manuscript_working_draft.md # Working paper skeleton for GitHub readers
docs/
DATA_CARD.md # Dataset scope, fields, release status, risks
ANNOTATION_PROTOCOL.md # Human validation plan and scoring dimensions
AI_WORKFLOW.md # AI-assisted research workflow and safeguards
SCORING_RUBRIC.md # Compact scoring rubric
MANUSCRIPT_STATUS.md # What is complete and what remains
data/
README.md
sample/legalbenchpro_cn_judgments_sample.csv
sample/legalbenchpro_public_exam_sample.csv
metadata/dataset_summary.json
metadata/model_configurations.csv
metadata/source_distribution.csv
scripts/
extract_public_sample.py # Rebuilds the public sample and metadata
render_research_snapshot.py # Renders the README overview figure
src/legalbenchpro/
workbook.py # Small workbook helpers used by scripts
tests/
test_workbook.py # Lightweight smoke tests for public utilities
The full workbook is not committed. To regenerate the public sample from a local private workbook:
python -m venv .venv
.\.venv\Scripts\Activate.ps1
pip install -r requirements.txt
$env:PYTHONPATH = "$PWD\src"
python .\scripts\extract_public_sample.py `
--workbook "C:\path\to\Data Set.xlsx" `
--out-dir data `
--cn-sample-size 10 `
--bar-sample-size 20 `
--max-cell-chars 420
python .\scripts\render_research_snapshot.pyThe repository includes a small test suite:
$env:PYTHONPATH = "$PWD\src"
python -m unittest discover -s tests
python -m compileall scripts srcThis repository is intentionally organized as a research-engineering artifact, not only as a dataset announcement. It demonstrates:
- Python scripts that regenerate public samples and metadata from a private workbook;
- explicit dataset documentation, release constraints, and annotation protocol files;
- lightweight tests for workbook parsing utilities;
- an audit trail for AI-assisted coding and research workflow decisions;
- manuscript-facing materials that separate current evidence from future validation.
This is a research preview, not a final benchmark release. The public content samples are excerpted and do not include the full prompt matrix, full reference answers, full model outputs, or human review sheets. The full dataset will require a final licensing, privacy, and source-distribution review before release.
Hongyu Wang. Manuscript and benchmark in preparation.
This repository is for research on model evaluation. It is not legal advice, a legal research product, or a substitute for jurisdiction-specific legal review.
