@botlearn/summarizer 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/LICENSE ADDED
@@ -0,0 +1,21 @@
1
+ MIT License
2
+
3
+ Copyright (c) 2025 BotLearn
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
package/README.md ADDED
@@ -0,0 +1,35 @@
1
+ # @botlearn/summarizer
2
+
3
+ > Content summarization with core argument extraction, key detail retention, and accuracy self-checking for OpenClaw Agent
4
+
5
+ ## Installation
6
+
7
+ ```bash
8
+ # via npm
9
+ npm install @botlearn/summarizer
10
+
11
+ # via clawhub
12
+ clawhub install @botlearn/summarizer
13
+ ```
14
+
15
+ ## Category
16
+
17
+ content-processing
18
+
19
+ ## Dependencies
20
+
21
+ None
22
+
23
+ ## Files
24
+
25
+ | File | Description |
26
+ |------|-------------|
27
+ | `manifest.json` | Skill metadata and configuration |
28
+ | `skill.md` | Role definition and activation rules |
29
+ | `knowledge/` | Domain knowledge documents |
30
+ | `strategies/` | Behavioral strategy definitions |
31
+ | `tests/` | Smoke and benchmark tests |
32
+
33
+ ## License
34
+
35
+ MIT
@@ -0,0 +1,76 @@
1
+ ---
2
+ domain: summarizer
3
+ topic: anti-patterns
4
+ priority: medium
5
+ ttl: 30d
6
+ ---
7
+
8
+ # Content Summarization — Anti-Patterns
9
+
10
+ ## Detail Loss Anti-Patterns
11
+
12
+ ### 1. Dropping Numbers and Dates
13
+ - **Problem**: Replacing specific quantitative data with vague language — e.g., summarizing "Revenue increased 34% to $4.2B in Q3 2024" as "Revenue increased significantly"
14
+ - **Impact**: The summary loses verifiability and decision-making value; "significantly" is subjective and unanchored
15
+ - **Fix**: Always carry forward key numbers, percentages, dates, and monetary values. If compression requires it, round numbers but never remove them. "$4.2B (up 34%, Q3 2024)" is acceptable compression; "revenue grew" is not
16
+
17
+ ### 2. Stripping Proper Nouns
18
+ - **Problem**: Replacing named entities with generic references — e.g., "The company announced" instead of "Anthropic announced"
19
+ - **Impact**: Loses attribution and specificity; makes the summary ambiguous when multiple entities are discussed
20
+ - **Fix**: Retain proper nouns for all entities central to the argument. Generic references are only acceptable for minor entities mentioned in passing
21
+
22
+ ### 3. Losing Temporal Context
23
+ - **Problem**: Omitting dates, time periods, or temporal qualifiers — e.g., "The study found correlation" instead of "A 2024 longitudinal study spanning 5 years found correlation"
24
+ - **Impact**: The reader cannot assess recency or relevance; conflates historical and current findings
25
+ - **Fix**: Include the year and study duration for research findings; include dates for events; include version numbers for technology
26
+
27
+ ## Structural Anti-Patterns
28
+
29
+ ### 4. Over-Generalization
30
+ - **Problem**: Collapsing specific, differentiated claims into a single vague statement — e.g., summarizing three distinct policy recommendations as "Several recommendations were made"
31
+ - **Impact**: Destroys the actionable specificity that makes the source valuable
32
+ - **Fix**: If the source lists N specific items and the summary can only fit M < N, select the M most important by impact or novelty and state them explicitly. Never replace a list with "several" or "various"
33
+
34
+ ### 5. Position Bias (Lead Bias)
35
+ - **Problem**: Over-representing content from the beginning of the document and under-representing content from the middle or end
36
+ - **Impact**: Conclusions, counter-arguments, and caveats (which typically appear later) are systematically omitted, producing a misleadingly one-sided summary
37
+ - **Fix**: Scan the entire document before summarizing. Use argument mapping to identify structurally important content regardless of position. Explicitly check: "Does my summary include content from the final third of the document?"
38
+
39
+ ### 6. Recency Bias (Tail Bias)
40
+ - **Problem**: The opposite of lead bias — over-representing the most recently read content because it is freshest in memory
41
+ - **Impact**: The opening thesis and early evidence are underweighted
42
+ - **Fix**: Build the argument map before writing the summary. Work from the map, not from memory of the reading order
43
+
44
+ ### 7. Flattening Argument Structure
45
+ - **Problem**: Presenting all claims at the same level, losing the hierarchical relationship between main arguments and supporting evidence
46
+ - **Impact**: The reader cannot distinguish the central thesis from supporting details; the summary reads as a list of disconnected facts
47
+ - **Fix**: Use the claim-evidence hierarchy from knowledge/domain.md. Lead with the thesis, support with top-level claims, and use indentation or explicit markers to show evidence supporting each claim
48
+
49
+ ## Accuracy Anti-Patterns
50
+
51
+ ### 8. Inference Injection
52
+ - **Problem**: Adding conclusions or causal claims that the source does not make — e.g., the source says "X and Y are correlated" but the summary says "X causes Y"
53
+ - **Impact**: The summary makes stronger claims than the source, potentially misleading the reader
54
+ - **Fix**: Preserve the source's hedging language. If the source says "suggests", "may", "correlates with", do not upgrade to "proves", "will", "causes". If you catch yourself adding a causal word not in the source, flag it
55
+
56
+ ### 9. Misattribution
57
+ - **Problem**: Attributing a claim to the wrong entity — e.g., attributing a critic's objection to the author, or attributing one researcher's finding to another
58
+ - **Impact**: Fundamentally changes the meaning; can misrepresent positions in a debate
59
+ - **Fix**: When the source contains multiple voices or perspectives, track attribution explicitly in the argument map. Verify each claim-attribution pair before including in the summary
60
+
61
+ ### 10. False Equivalence in Multi-Document Summaries
62
+ - **Problem**: Giving equal weight to a well-evidenced majority view and a poorly supported minority view, or vice versa
63
+ - **Impact**: Distorts the landscape of expert consensus; can give fringe positions undue prominence
64
+ - **Fix**: Weight claims by evidence quality (see argument strength in knowledge/domain.md). If 8 sources agree and 1 disagrees, state the consensus and note the dissent — do not present them as equally supported
65
+
66
+ ## Output Anti-Patterns
67
+
68
+ ### 11. Wall-of-Text Summaries
69
+ - **Problem**: Producing a summary as a single dense paragraph with no structural markers
70
+ - **Impact**: Defeats the purpose of summarization; the reader must re-read to find specific points
71
+ - **Fix**: Use structural formatting — bullet points, numbered lists, or short paragraphs with clear topic sentences. Match the output structure to the user's likely use case (scanning vs. reading)
72
+
73
+ ### 12. Summary Longer Than Necessary
74
+ - **Problem**: Including filler phrases ("It is worth noting that", "In conclusion, it can be said that") or restating the same point in multiple ways
75
+ - **Impact**: Inflates the summary length without adding information; wastes the reader's time
76
+ - **Fix**: After drafting, apply a self-edit pass: can any sentence be removed without losing information? If yes, remove it. Target zero-filler summaries
@@ -0,0 +1,130 @@
1
+ ---
2
+ domain: summarizer
3
+ topic: extractive-abstractive-techniques-and-detail-preservation
4
+ priority: high
5
+ ttl: 30d
6
+ ---
7
+
8
+ # Content Summarization — Best Practices
9
+
10
+ ## Extractive vs. Abstractive Summarization
11
+
12
+ ### Extractive Summarization
13
+ Selects and assembles existing sentences or phrases from the source without modification.
14
+
15
+ **When to use extractive**:
16
+ - Legal, medical, or regulatory documents where precise wording carries meaning
17
+ - When the user explicitly requests "in the author's own words"
18
+ - When quantitative data is dense and paraphrasing risks numerical errors
19
+ - For attributing specific claims — direct quotes preserve accountability
20
+
21
+ **Technique**:
22
+ 1. Score each sentence by information density (see knowledge/domain.md)
23
+ 2. Select top-N sentences by density score, preserving document order
24
+ 3. Add minimal connective tissue between extracted sentences for readability
25
+ 4. Indicate extracted passages with source markers (e.g., [para 3], [section 2.1])
26
+
27
+ ### Abstractive Summarization
28
+ Generates new sentences that capture the meaning of the source in compressed form.
29
+
30
+ **When to use abstractive**:
31
+ - General-purpose summaries where readability is the priority
32
+ - When compression ratio exceeds 5:1 (extractive becomes choppy at high compression)
33
+ - When synthesizing multiple documents into a unified summary
34
+ - When the source is poorly written and extraction would perpetuate unclear language
35
+
36
+ **Technique**:
37
+ 1. Build an argument map (see knowledge/domain.md) to capture logical structure
38
+ 2. Rewrite each branch of the argument map as a concise statement
39
+ 3. Verify every factual claim in the rewritten text against the source
40
+ 4. Flag any inference you made that goes beyond the source material
41
+
42
+ ### Hybrid Approach (Recommended Default)
43
+ Combine extractive and abstractive methods for optimal accuracy and readability:
44
+ - **Abstractive** for narrative flow, transitions, and structural framing
45
+ - **Extractive** for quantitative data, direct quotes, and precise technical claims
46
+ - Mark the boundary: use quotation marks or source markers for extracted content
47
+
48
+ ## Detail Preservation Framework
49
+
50
+ ### Tier 1 — Must Preserve (Never Omit)
51
+ These details, if present in the source, must appear in any summary regardless of length:
52
+ - **Central thesis or conclusion** — The main point of the document
53
+ - **Quantitative results** — Key numbers, percentages, financial figures, dates
54
+ - **Named entities driving the narrative** — People, organizations, products central to the argument
55
+ - **Causal claims with evidence** — "X caused Y" where Y is a significant outcome
56
+ - **Contradictions or caveats** — If the source acknowledges limitations, the summary must too
57
+
58
+ ### Tier 2 — Preserve When Space Permits
59
+ - Supporting examples (keep the strongest one per claim)
60
+ - Methodological details (sample size, data sources, time period)
61
+ - Secondary claims that reinforce but do not introduce new information
62
+ - Historical context that frames the current discussion
63
+
64
+ ### Tier 3 — Safe to Omit
65
+ - Repeated examples after the first strong one
66
+ - Author biography or credentials (unless relevant to credibility assessment)
67
+ - Acknowledgments, funding disclosures, boilerplate
68
+ - Detailed literature reviews (summarize as "building on prior work by [key names]")
69
+
70
+ ## Multi-Document Summarization
71
+
72
+ When summarizing across multiple sources, additional practices apply:
73
+
74
+ ### 1. Claim Reconciliation
75
+ - Identify claims that appear across multiple sources → strengthen confidence
76
+ - Identify contradictory claims → present both with attribution
77
+ - Identify claims unique to a single source → note as "according to [source]"
78
+
79
+ ### 2. Overlap Deduplication
80
+ - When multiple sources state the same fact, consolidate into one statement
81
+ - Cite the most authoritative source for the consolidated claim
82
+ - Note the number of corroborating sources if relevant (e.g., "confirmed by 3 independent studies")
83
+
84
+ ### 3. Gap Analysis
85
+ - After merging, check: are there aspects of the topic that no source covers?
86
+ - Flag gaps explicitly: "None of the sources address [aspect]"
87
+ - This prevents the user from assuming comprehensive coverage when gaps exist
88
+
89
+ ### 4. Synthesis Structure
90
+ For multi-document summaries, organize by theme rather than by source:
91
+
92
+ ```
93
+ Theme A
94
+ ├── Source 1 finding
95
+ ├── Source 2 finding (corroborates)
96
+ └── Source 3 finding (contradicts — note discrepancy)
97
+
98
+ Theme B
99
+ ├── Source 1 finding (only source)
100
+ └── [Gap: not addressed by other sources]
101
+ ```
102
+
103
+ ## Compression Ratio Guidelines
104
+
105
+ | User Request | Target Ratio | Strategy |
106
+ |-------------|-------------|----------|
107
+ | "TLDR" / "one-liner" | 20:1 or higher | Central thesis + single most important finding |
108
+ | "Key points" / "brief summary" | 10:1 | Thesis + top 3-5 claims with headline evidence |
109
+ | "Summary" (default) | 5:1 | Full argument map preserved; Tier 1 + select Tier 2 details |
110
+ | "Detailed summary" | 3:1 | All Tier 1 + Tier 2 details; methodology included |
111
+ | "Executive briefing" | 5:1 | Thesis + implications + recommended actions; omit methodology |
112
+
113
+ ## Output Formatting Best Practices
114
+
115
+ ### For Structured Output
116
+ - Lead with a one-sentence thesis summary
117
+ - Use bullet points for key findings (one bullet per claim)
118
+ - Group related findings under thematic headings
119
+ - End with implications or open questions
120
+
121
+ ### For Prose Output
122
+ - Open with the central finding or conclusion (inverted pyramid)
123
+ - Support with evidence in descending order of importance
124
+ - Close with caveats, limitations, or open questions
125
+ - Keep paragraphs to 3-4 sentences maximum
126
+
127
+ ### Source Traceability
128
+ - Annotate key claims with source references: [Source: section/paragraph]
129
+ - For multi-document summaries, attribute each claim to its source
130
+ - This enables the user to verify any claim against the original material
@@ -0,0 +1,114 @@
1
+ ---
2
+ domain: summarizer
3
+ topic: discourse-structure-and-argument-mapping
4
+ priority: high
5
+ ttl: 30d
6
+ ---
7
+
8
+ # Content Summarization — Discourse Types, Argument Mapping & Information Density
9
+
10
+ ## Discourse Types
11
+
12
+ Understanding the discourse type of a source document is the foundation of effective summarization. Each type has a distinct internal structure that dictates where core information resides and how it should be compressed.
13
+
14
+ ### 1. Argumentative Discourse
15
+ - **Structure**: Thesis → Supporting claims → Evidence → Counter-arguments → Conclusion
16
+ - **Key signal words**: "therefore", "however", "because", "despite", "in contrast", "evidence suggests"
17
+ - **Summarization focus**: Preserve the thesis, the strongest supporting claims, key evidence (especially quantitative), and any concessions or counter-arguments acknowledged by the author
18
+ - **Compression target**: Thesis and top 2-3 supporting claims with their strongest evidence
19
+
20
+ ### 2. Expository Discourse
21
+ - **Structure**: Topic introduction → Categorical breakdown → Details per category → Synthesis
22
+ - **Key signal words**: "first", "second", "in addition", "for example", "such as", "specifically"
23
+ - **Summarization focus**: Preserve the categorical structure; for each category, retain the defining characteristic and one concrete example
24
+ - **Compression target**: Topic + category headings + key differentiator per category
25
+
26
+ ### 3. Narrative Discourse
27
+ - **Structure**: Setting → Rising action → Climax → Falling action → Resolution
28
+ - **Key signal words**: "then", "next", "suddenly", "as a result", "finally", "meanwhile"
29
+ - **Summarization focus**: Preserve the causal chain — what happened, why, and what resulted; retain key actors and turning points
30
+ - **Compression target**: Who + initiating event + key turning points + outcome
31
+
32
+ ### 4. Descriptive Discourse
33
+ - **Structure**: Overview → Feature-by-feature detail → Comparative context
34
+ - **Key signal words**: "characterized by", "consists of", "appears as", "resembles", "differs from"
35
+ - **Summarization focus**: Capture the entity's defining features and any comparative positioning; omit redundant descriptive detail
36
+ - **Compression target**: Entity + 3-5 distinguishing features + comparative positioning
37
+
38
+ ### 5. Procedural Discourse
39
+ - **Structure**: Goal statement → Prerequisites → Sequential steps → Verification
40
+ - **Key signal words**: "first", "then", "next", "ensure that", "before", "after", "warning"
41
+ - **Summarization focus**: Preserve the goal, critical prerequisites, step order, and any warnings or failure conditions; minor sub-steps can be collapsed
42
+ - **Compression target**: Goal + prerequisites + collapsed step sequence + critical warnings
43
+
44
+ ## Argument Mapping
45
+
46
+ Argument mapping is the process of identifying the logical structure within a document. This is critical for producing summaries that preserve the author's reasoning rather than just surface-level facts.
47
+
48
+ ### Claim-Evidence Hierarchy
49
+
50
+ ```
51
+ Central Thesis
52
+ ├── Claim 1 (supports thesis)
53
+ │ ├── Evidence 1a (data, statistic, citation)
54
+ │ ├── Evidence 1b (example, case study)
55
+ │ └── Qualifier (conditions under which claim holds)
56
+ ├── Claim 2 (supports thesis)
57
+ │ ├── Evidence 2a
58
+ │ └── Counter-evidence (acknowledged weakness)
59
+ ├── Counter-Argument (opposing view)
60
+ │ ├── Evidence for counter-argument
61
+ │ └── Author's rebuttal
62
+ └── Conclusion (restated thesis + implications)
63
+ ```
64
+
65
+ ### Identifying Claims vs. Evidence
66
+
67
+ | Type | Characteristics | Examples |
68
+ |------|----------------|---------|
69
+ | **Claim** | Assertive, debatable, evaluative | "Remote work increases productivity", "Policy X is ineffective" |
70
+ | **Evidence** | Factual, verifiable, specific | "A 2024 Stanford study found 13% productivity increase", "$2.3 billion in losses" |
71
+ | **Qualifier** | Conditional, limiting scope | "In knowledge-work sectors", "When implemented with proper tooling" |
72
+ | **Warrant** | Implicit assumption linking evidence to claim | "Because productivity correlates with employee satisfaction" |
73
+
74
+ ### Argument Strength Assessment
75
+
76
+ When multiple claims compete for inclusion in a summary, prioritize by argument strength:
77
+
78
+ 1. **Strong**: Claim + multiple independent evidence sources + acknowledged qualifiers
79
+ 2. **Moderate**: Claim + single evidence source or claim + logical reasoning without empirical data
80
+ 3. **Weak**: Claim with no evidence, appeal to authority alone, or anecdotal support
81
+ 4. **Contested**: Claim with counter-evidence presented — must include both sides
82
+
83
+ ## Information Density Analysis
84
+
85
+ Not all sentences carry equal information value. Information density determines which content survives compression.
86
+
87
+ ### High-Density Signals (Preserve)
88
+ - **Quantitative data**: Percentages, dollar amounts, dates, counts, measurements
89
+ - Example: "Revenue grew 34% year-over-year to $4.2B in Q3 2024"
90
+ - **Proper nouns**: People, organizations, places, product names
91
+ - Example: "CEO Maria Chen announced the partnership with Anthropic"
92
+ - **Causal relationships**: Explicit cause-effect statements
93
+ - Example: "The tariff increase led to a 15% drop in imports within 6 months"
94
+ - **Novel claims**: Information the reader likely does not already know
95
+ - Example: "Contrary to prior research, the study found no correlation between X and Y"
96
+ - **Definitions**: First introduction of a key term or concept
97
+ - Example: "Retrieval-augmented generation (RAG) combines a retriever with a generator"
98
+
99
+ ### Low-Density Signals (Compress or Remove)
100
+ - **Redundant restatements**: Same idea expressed in different words
101
+ - **Background context**: Common knowledge or widely known facts
102
+ - **Hedging language**: "It is important to note that", "One might argue"
103
+ - **Transitional padding**: "Having examined X, let us now turn to Y"
104
+ - **Excessive examples**: When one example suffices to illustrate a point, additional examples can be dropped
105
+
106
+ ### Density Scoring Heuristic
107
+
108
+ For each sentence, assign a density score (0-3):
109
+ - **3**: Contains quantitative data + causal claim + proper noun
110
+ - **2**: Contains a novel claim or key evidence
111
+ - **1**: Provides context or elaboration on a higher-density sentence
112
+ - **0**: Transitional, redundant, or common knowledge
113
+
114
+ Sentences scoring 2-3 are primary candidates for summary inclusion. Sentences scoring 1 are included only if the summary length permits. Sentences scoring 0 are excluded.
package/manifest.json ADDED
@@ -0,0 +1,26 @@
1
+ {
2
+ "name": "@botlearn/summarizer",
3
+ "version": "0.1.0",
4
+ "description": "Content summarization with core argument extraction, key detail retention, and accuracy self-checking for OpenClaw Agent",
5
+ "category": "content-processing",
6
+ "author": "BotLearn",
7
+ "benchmarkDimension": "content-understanding",
8
+ "expectedImprovement": 45,
9
+ "dependencies": {},
10
+ "compatibility": {
11
+ "openclaw": ">=0.5.0"
12
+ },
13
+ "files": {
14
+ "skill": "skill.md",
15
+ "knowledge": [
16
+ "knowledge/domain.md",
17
+ "knowledge/best-practices.md",
18
+ "knowledge/anti-patterns.md"
19
+ ],
20
+ "strategies": [
21
+ "strategies/main.md"
22
+ ],
23
+ "smokeTest": "tests/smoke.json",
24
+ "benchmark": "tests/benchmark.json"
25
+ }
26
+ }
package/package.json ADDED
@@ -0,0 +1,35 @@
1
+ {
2
+ "name": "@botlearn/summarizer",
3
+ "version": "0.1.0",
4
+ "description": "Content summarization with core argument extraction, key detail retention, and accuracy self-checking for OpenClaw Agent",
5
+ "type": "module",
6
+ "main": "manifest.json",
7
+ "files": [
8
+ "manifest.json",
9
+ "skill.md",
10
+ "knowledge/",
11
+ "strategies/",
12
+ "tests/",
13
+ "README.md"
14
+ ],
15
+ "keywords": [
16
+ "botlearn",
17
+ "openclaw",
18
+ "skill",
19
+ "content-processing"
20
+ ],
21
+ "author": "BotLearn",
22
+ "license": "MIT",
23
+ "repository": {
24
+ "type": "git",
25
+ "url": "https://github.com/readai-team/botlearn-awesome-skills.git",
26
+ "directory": "packages/skills/summarizer"
27
+ },
28
+ "homepage": "https://github.com/readai-team/botlearn-awesome-skills/tree/main/packages/skills/summarizer",
29
+ "bugs": {
30
+ "url": "https://github.com/readai-team/botlearn-awesome-skills/issues"
31
+ },
32
+ "publishConfig": {
33
+ "access": "public"
34
+ }
35
+ }
package/skill.md ADDED
@@ -0,0 +1,43 @@
1
+ ---
2
+ name: summarizer
3
+ role: Content Summarization Specialist
4
+ version: 1.0.0
5
+ triggers:
6
+ - "summarize"
7
+ - "summary"
8
+ - "key points"
9
+ - "TLDR"
10
+ - "digest"
11
+ - "main ideas"
12
+ - "boil down"
13
+ ---
14
+
15
+ # Role
16
+
17
+ You are a Content Summarization Specialist. When activated, you analyze documents, articles, and multi-source content to extract core arguments, retain critical details (numbers, dates, names, causal claims), and produce accurate, well-structured summaries calibrated to the user's desired length and depth.
18
+
19
+ # Capabilities
20
+
21
+ 1. Identify discourse structure (narrative, argumentative, expository, descriptive, procedural) and adapt extraction strategy accordingly
22
+ 2. Extract and map core arguments, distinguishing claims from evidence, and preserving the logical chain between them
23
+ 3. Prioritize details by information density — retain quantitative data, proper nouns, causal relationships, and novel insights while compressing redundant or low-signal passages
24
+ 4. Synthesize multi-document inputs into unified summaries, reconciling overlapping claims and surfacing contradictions
25
+ 5. Perform accuracy self-checks by verifying that every key claim in the summary traces back to a specific passage in the source material
26
+
27
+ # Constraints
28
+
29
+ 1. Never fabricate details — every fact, number, date, and name in the summary must appear in the source material
30
+ 2. Never omit quantitative data (percentages, dollar amounts, dates, counts) that support core arguments
31
+ 3. Never flatten nuance — if the source presents competing viewpoints, the summary must reflect that tension
32
+ 4. Always preserve attribution — if a claim is attributed to a specific person or organization, maintain that attribution
33
+ 5. Always state the summary's compression ratio and flag if the requested length risks losing critical information
34
+
35
+ # Activation
36
+
37
+ WHEN the user requests a summary, key points, TLDR, or digest of content:
38
+ 1. Determine target length and depth (brief / standard / detailed) from user cues
39
+ 2. Analyze document structure following strategies/main.md
40
+ 3. Apply discourse type recognition from knowledge/domain.md
41
+ 4. Extract arguments and details using knowledge/best-practices.md
42
+ 5. Verify against knowledge/anti-patterns.md to avoid common summarization errors
43
+ 6. Output a structured summary with source traceability annotations
@@ -0,0 +1,85 @@
1
+ ---
2
+ strategy: summarizer
3
+ version: 1.0.0
4
+ steps: 5
5
+ ---
6
+
7
+ # Content Summarization Strategy
8
+
9
+ ## Step 1: Structure Identification
10
+ - Read the full source material before beginning summarization — never summarize while reading for the first time
11
+ - Classify the discourse type: argumentative / expository / narrative / descriptive / procedural (see knowledge/domain.md)
12
+ - Identify the document's structural skeleton:
13
+ - IF argumentative THEN locate thesis, claims, evidence blocks, counter-arguments, conclusion
14
+ - IF expository THEN locate topic statement, category headings, detail sections, synthesis
15
+ - IF narrative THEN locate setting, key actors, initiating event, turning points, resolution
16
+ - IF descriptive THEN locate subject overview, feature enumerations, comparative sections
17
+ - IF procedural THEN locate goal statement, prerequisites, step sequence, warnings, verification
18
+ - Determine the total length and estimate a target compression ratio based on user intent:
19
+ - "TLDR" / "one-liner" → 20:1+
20
+ - "key points" / "brief" → 10:1
21
+ - "summary" (default) → 5:1
22
+ - "detailed summary" → 3:1
23
+ - IF the source is multi-document THEN note the number of documents and scan for thematic overlap before proceeding
24
+
25
+ ## Step 2: Argument Extraction
26
+ - Build a claim-evidence hierarchy using the argument mapping framework from knowledge/domain.md:
27
+ - Central thesis → supporting claims → evidence for each claim → qualifiers and counter-arguments
28
+ - For each claim identified, record:
29
+ - The claim statement
30
+ - The type and strength of its evidence (quantitative data, case study, expert opinion, logical reasoning)
31
+ - Any qualifiers or conditions limiting the claim's scope
32
+ - Any counter-evidence or objections acknowledged by the source
33
+ - Rank claims by argument strength:
34
+ - **Strong**: Multiple independent evidence sources + qualifiers acknowledged
35
+ - **Moderate**: Single evidence source or logical reasoning only
36
+ - **Weak**: Assertion without evidence or anecdotal support only
37
+ - **Contested**: Counter-evidence presented — both sides must appear in summary
38
+ - IF the source is multi-document THEN merge argument maps:
39
+ - Corroborated claims (multiple sources) → high confidence
40
+ - Contradictory claims → flag with attribution to each source
41
+ - Unique claims (single source) → note as "according to [source]"
42
+
43
+ ## Step 3: Detail Prioritization
44
+ - Score each extracted detail by information density (see knowledge/domain.md):
45
+ - **Score 3**: Quantitative data + causal claim + named entity (always include)
46
+ - **Score 2**: Novel claim or key evidence (include in standard+ summaries)
47
+ - **Score 1**: Contextual elaboration (include only in detailed summaries)
48
+ - **Score 0**: Transitional, redundant, or common knowledge (exclude)
49
+ - Apply the detail preservation tiers from knowledge/best-practices.md:
50
+ - **Tier 1 — Must Preserve**: Central thesis, quantitative results, key named entities, causal claims with evidence, contradictions and caveats
51
+ - **Tier 2 — Space Permitting**: Strongest example per claim, methodology details, secondary reinforcing claims, historical context
52
+ - **Tier 3 — Safe to Omit**: Repeated examples, author credentials, acknowledgments, detailed literature reviews
53
+ - Cross-check against anti-patterns from knowledge/anti-patterns.md:
54
+ - Am I dropping any numbers or dates? → Restore them
55
+ - Am I stripping proper nouns? → Restore attribution
56
+ - Am I over-generalizing specific lists? → Name the top items explicitly
57
+ - Am I showing position bias? → Verify coverage spans the full document
58
+
59
+ ## Step 4: Synthesis
60
+ - Construct the summary using the hybrid approach from knowledge/best-practices.md:
61
+ - **Abstractive**: Use for narrative flow, transitions, and structural framing
62
+ - **Extractive**: Use for quantitative data, direct quotes, and precise technical claims (mark with source references)
63
+ - Structure the output based on user intent:
64
+ - IF "TLDR" THEN one sentence: thesis + most important finding
65
+ - IF "key points" THEN bullet list: thesis + top 3-5 claims with headline evidence
66
+ - IF "summary" THEN structured prose: full argument skeleton with Tier 1 details
67
+ - IF "detailed summary" THEN section-by-section: all Tier 1 + Tier 2 details with source annotations
68
+ - IF multi-document THEN organize by theme, not by source
69
+ - Apply formatting best practices:
70
+ - Lead with the central conclusion (inverted pyramid)
71
+ - One idea per bullet point or paragraph
72
+ - Use headings or numbered lists for multi-part summaries
73
+ - Add source traceability markers for key claims
74
+
75
+ ## Step 5: Accuracy Self-Check
76
+ - Perform a systematic verification pass before outputting the summary:
77
+ 1. **Fact check**: For every number, date, percentage, and proper noun in the summary, verify it appears in the source material — if it does not, remove or correct it
78
+ 2. **Claim fidelity**: For every causal or evaluative statement, verify the source makes the same claim at the same strength — if the summary says "causes" but the source says "correlates with", downgrade the language
79
+ 3. **Attribution check**: For every attributed claim ("X said", "according to Y"), verify the attribution is correct and not confused with another entity
80
+ 4. **Completeness check**: Review the argument map — is any strong or contested claim missing from the summary? If yes and space permits, add it
81
+ 5. **Position bias check**: Does the summary draw from all sections of the source, or only the beginning? If the final third is unrepresented, reassess
82
+ 6. **Anti-pattern scan**: Run through the 12 anti-patterns in knowledge/anti-patterns.md — does the summary violate any?
83
+ - IF any check fails THEN revise the summary and re-run the verification
84
+ - State the compression ratio in the output: "[Original: ~X words → Summary: ~Y words, compression ratio Z:1]"
85
+ - IF the target compression risks losing Tier 1 details THEN warn the user: "Note: The requested length requires omitting [specific detail]. Expand to [length] for full coverage."
@@ -0,0 +1,476 @@
1
+ {
2
+ "version": "0.0.1",
3
+ "dimension": "content-understanding",
4
+ "tasks": [
5
+ {
6
+ "id": "bench-easy-01",
7
+ "difficulty": "easy",
8
+ "description": "Summarize a short factual news article with key statistics",
9
+ "input": "Summarize the following news article:\n\n---\nGlobal Renewable Energy Hits Record in 2024\n\nThe International Energy Agency (IEA) reported on January 15, 2025 that global renewable energy capacity additions reached 507 gigawatts (GW) in 2024, a 25% increase from the 405 GW added in 2023. Solar photovoltaic (PV) accounted for 75% of new additions at 380 GW, while wind power contributed 117 GW.\n\nChina led installations with 292 GW (57% of global total), followed by the United States at 52 GW, India at 29 GW, and Germany at 18 GW. The IEA's Executive Director Fatih Birol stated that renewables are now 'the cheapest source of new electricity in virtually every country.'\n\nDespite the record additions, the IEA warned that the pace must accelerate to 1,000 GW annually by 2030 to meet Paris Agreement targets. Grid infrastructure remains a bottleneck, with approximately 1,500 GW of renewable projects globally awaiting grid connection.\n---",
10
+ "rubric": [
11
+ {
12
+ "criterion": "Key Fact Retention",
13
+ "weight": 0.4,
14
+ "scoring": {
15
+ "5": "Retains: 507 GW total, 25% increase, solar 380 GW/wind 117 GW split, China 292 GW, 1,000 GW 2030 target, 1,500 GW backlog",
16
+ "3": "Retains total and top 2-3 statistics but misses some",
17
+ "1": "Only retains 1-2 numbers",
18
+ "0": "No quantitative data preserved"
19
+ }
20
+ },
21
+ {
22
+ "criterion": "Structure Accuracy",
23
+ "weight": 0.3,
24
+ "scoring": {
25
+ "5": "Captures the three-part structure: record achievement, country breakdown, and gap/warning",
26
+ "3": "Captures achievement and one other aspect",
27
+ "1": "Only captures the headline finding",
28
+ "0": "Misrepresents the article structure"
29
+ }
30
+ },
31
+ {
32
+ "criterion": "Conciseness",
33
+ "weight": 0.3,
34
+ "scoring": {
35
+ "5": "Summary is 40-80 words, well-compressed, no filler",
36
+ "3": "Summary is reasonable length but contains some unnecessary detail",
37
+ "1": "Summary is nearly as long as the original or extremely truncated",
38
+ "0": "No meaningful compression"
39
+ }
40
+ }
41
+ ],
42
+ "expectedScoreWithout": 45,
43
+ "expectedScoreWith": 85
44
+ },
45
+ {
46
+ "id": "bench-easy-02",
47
+ "difficulty": "easy",
48
+ "description": "Extract key points from a procedural/instructional text",
49
+ "input": "Give me the key points from this guide:\n\n---\nSetting Up a Production Kubernetes Cluster: Essential Steps\n\n1. Capacity Planning: Before provisioning, calculate expected workload. A typical starting configuration for a medium SaaS application is 3 control plane nodes (4 vCPU, 16 GB RAM each) and 5-10 worker nodes (8 vCPU, 32 GB RAM each). Over-provisioning by 30% is recommended for burst capacity.\n\n2. Network Configuration: Use a CNI plugin (Calico or Cilium recommended). Ensure pod CIDR does not overlap with existing VPC ranges. Allocate at least a /16 for pod networking (65,536 addresses). Enable network policies from day one — retrofitting is significantly harder.\n\n3. Security Hardening: Enable RBAC (mandatory since Kubernetes 1.6). Disable anonymous authentication. Use Pod Security Admission (replacing PodSecurityPolicy since v1.25). Rotate certificates every 90 days. Store secrets in an external manager (HashiCorp Vault or AWS Secrets Manager), not in etcd.\n\n4. Monitoring Stack: Deploy Prometheus + Grafana for metrics, Loki for logs, and Jaeger for distributed tracing. Set up alerts for: node CPU > 80%, pod restart count > 3 in 5 minutes, and persistent volume usage > 85%.\n\n5. Disaster Recovery: Back up etcd every 6 hours using etcdctl snapshot. Test restoration quarterly. Maintain infrastructure-as-code (Terraform or Pulumi) so the entire cluster can be recreated from scratch in under 2 hours. Document the runbook.\n---",
50
+ "rubric": [
51
+ {
52
+ "criterion": "Completeness",
53
+ "weight": 0.4,
54
+ "scoring": {
55
+ "5": "All 5 sections represented with their most critical actionable detail from each",
56
+ "3": "4 of 5 sections represented or some sections lack specifics",
57
+ "1": "Only 2-3 sections captured",
58
+ "0": "Major sections missing"
59
+ }
60
+ },
61
+ {
62
+ "criterion": "Technical Precision",
63
+ "weight": 0.3,
64
+ "scoring": {
65
+ "5": "Preserves specific technical details: node specs, CIDR range, tool names (Calico/Cilium, Vault), alert thresholds, backup intervals",
66
+ "3": "Preserves tool names but loses specific configurations",
67
+ "1": "Generic summary without technical specifics",
68
+ "0": "Technical details incorrect or fabricated"
69
+ }
70
+ },
71
+ {
72
+ "criterion": "Format Appropriateness",
73
+ "weight": 0.3,
74
+ "scoring": {
75
+ "5": "Uses bullet or numbered list matching the procedural source; each point is concise and actionable",
76
+ "3": "Structured but some points are verbose or unclear",
77
+ "1": "Prose format that buries the procedural structure",
78
+ "0": "Unstructured output"
79
+ }
80
+ }
81
+ ],
82
+ "expectedScoreWithout": 40,
83
+ "expectedScoreWith": 82
84
+ },
85
+ {
86
+ "id": "bench-easy-03",
87
+ "difficulty": "easy",
88
+ "description": "TLDR-style ultra-brief summary of a product announcement",
89
+ "input": "TLDR this announcement:\n\n---\nAnthropic Launches Claude 4 with Expanded Capabilities\n\nSan Francisco, March 15, 2025 — Anthropic today announced Claude 4, the latest version of its AI assistant. The new model features a 500,000-token context window (up from 200,000 in Claude 3.5), improved reasoning capabilities scoring 92% on the GPQA benchmark (up from 65%), and native multimodal support for images, PDFs, and audio.\n\nKey pricing: Claude 4 Opus is available at $15 per million input tokens and $75 per million output tokens through the API. A new 'Projects' feature allows enterprise users to create persistent knowledge bases that Claude can reference across conversations.\n\nThe model was trained using Anthropic's Constitutional AI 2.0 framework, which the company claims reduces harmful outputs by 60% compared to the previous version while maintaining helpfulness scores. Claude 4 is available immediately through the API and web interface, with mobile apps launching in April 2025.\n---",
90
+ "rubric": [
91
+ {
92
+ "criterion": "Brevity",
93
+ "weight": 0.35,
94
+ "scoring": {
95
+ "5": "Summary is 1-2 sentences (under 40 words) capturing the essential announcement",
96
+ "3": "Summary is 3-4 sentences, slightly verbose for a TLDR",
97
+ "1": "Summary exceeds 5 sentences",
98
+ "0": "No meaningful compression"
99
+ }
100
+ },
101
+ {
102
+ "criterion": "Essential Information",
103
+ "weight": 0.4,
104
+ "scoring": {
105
+ "5": "Captures: what (Claude 4), who (Anthropic), and at least 2 key improvements (context window, GPQA score, multimodal, or pricing)",
106
+ "3": "Captures product and company but only 1 specific improvement",
107
+ "1": "Generic statement without specifics",
108
+ "0": "Misidentifies the product or company"
109
+ }
110
+ },
111
+ {
112
+ "criterion": "Accuracy",
113
+ "weight": 0.25,
114
+ "scoring": {
115
+ "5": "All stated facts match the source exactly — no number errors or misattributions",
116
+ "3": "Minor imprecision (e.g., rounded numbers that were already round)",
117
+ "1": "Contains a factual error",
118
+ "0": "Multiple factual errors or fabricated details"
119
+ }
120
+ }
121
+ ],
122
+ "expectedScoreWithout": 45,
123
+ "expectedScoreWith": 88
124
+ },
125
+ {
126
+ "id": "bench-med-01",
127
+ "difficulty": "medium",
128
+ "description": "Summarize an argumentative text with competing viewpoints and nuance",
129
+ "input": "Summarize the following policy analysis, preserving the competing viewpoints:\n\n---\nThe Universal Basic Income Debate: Evidence from Recent Pilots\n\nThe concept of Universal Basic Income (UBI) has moved from theoretical discussion to empirical testing, with over 50 pilot programs conducted worldwide since 2017. The evidence is both encouraging and cautionary, with results varying significantly based on program design, cultural context, and measurement methodology.\n\nThe strongest evidence for UBI comes from Finland's 2017-2018 experiment, which provided 2,000 unemployed citizens with EUR 560/month unconditionally. The final report, published by Kela in 2020, found that recipients experienced a 0.5 percentage point increase in employment compared to the control group, a statistically significant but modest effect. More notably, recipients reported substantially higher life satisfaction (7.3 vs 6.8 on a 10-point scale) and lower stress levels, with a 37% reduction in self-reported depression symptoms.\n\nStockton, California's SEED program (2019-2021) provided 125 residents with $500/month. UC Berkeley researchers found that full-time employment among recipients rose from 28% to 40%, contradicting the hypothesis that cash transfers reduce work incentive. However, critics note the small sample size and the presence of a Hawthorne effect.\n\nSkeptical perspectives come from several directions. Economist Gregory Mankiw argues that UBI at a meaningful level ($1,000/month per adult in the US) would cost approximately $3.1 trillion annually — roughly 75% of the current federal budget — making it fiscally unsustainable without massive tax increases or program consolidation. The Congressional Budget Office estimated that funding UBI through income tax alone would require a flat rate increase of 39 percentage points.\n\nA different criticism comes from labor economists like Daron Acemoglu, who contends that UBI addresses the symptom (income insufficiency) rather than the cause (labor market disruption from automation). Acemoglu advocates instead for wage subsidies and retraining programs that maintain the dignity and social integration benefits of employment.\n\nThe Kenya GiveDirectly study, the largest randomized controlled trial of basic income (20,000 recipients, 12-year program), has published interim results showing a 27% increase in household assets and a 13% increase in small business revenues among recipients. However, the study also found that recipients in villages where only some people received transfers experienced increased social tension, raising questions about targeting versus universality.\n\nProponents like economist Guy Standing argue that UBI should not be evaluated solely as an employment intervention but as a freedom-enhancing policy that gives individuals bargaining power and the ability to refuse exploitative work. Standing points to India's Madhya Pradesh pilot, where UBI recipients, particularly women, showed increased likelihood of starting small businesses (32% vs 19% in control) and girls' school attendance increased by 25%.\n\nThe emerging view among policy researchers is that UBI works best as a complement to, not a replacement for, existing social safety nets — a position articulated by Stanford's Basic Income Lab. The optimal design appears to involve smaller, unconditional payments ($300-500/month) paired with maintained access to healthcare, housing assistance, and education support.\n---",
130
+ "rubric": [
131
+ {
132
+ "criterion": "Viewpoint Balance",
133
+ "weight": 0.3,
134
+ "scoring": {
135
+ "5": "Presents all major perspectives: pro-UBI evidence (Finland, Stockton, Kenya, India), fiscal criticism (Mankiw/CBO), structural criticism (Acemoglu), and the emerging synthesis position — without favoring any",
136
+ "3": "Covers pro and con but misses 1-2 perspectives or shows implicit bias",
137
+ "1": "Presents mostly one side, mentions opposition superficially",
138
+ "0": "One-sided summary or misrepresents positions"
139
+ }
140
+ },
141
+ {
142
+ "criterion": "Quantitative Fidelity",
143
+ "weight": 0.25,
144
+ "scoring": {
145
+ "5": "Retains key numbers from both sides: EUR 560/month, employment effect (0.5pp or 28%→40%), $3.1T cost, 39pp tax increase, 27% asset increase, 32% vs 19% business starts",
146
+ "3": "Retains 4-5 key statistics with correct attribution",
147
+ "1": "Retains 1-2 numbers or uses vague quantifiers",
148
+ "0": "No numbers preserved or numbers are incorrect"
149
+ }
150
+ },
151
+ {
152
+ "criterion": "Nuance Preservation",
153
+ "weight": 0.25,
154
+ "scoring": {
155
+ "5": "Captures qualifications: Finland effect was 'modest', Stockton had small sample/Hawthorne, Kenya showed social tension, complement-not-replacement conclusion",
156
+ "3": "Captures 2-3 qualifications",
157
+ "1": "Presents findings as unqualified",
158
+ "0": "Strips all nuance, presents black-and-white conclusions"
159
+ }
160
+ },
161
+ {
162
+ "criterion": "Structural Coherence",
163
+ "weight": 0.2,
164
+ "scoring": {
165
+ "5": "Organized thematically (evidence → criticisms → synthesis) with clear flow; appropriate length (150-250 words)",
166
+ "3": "Logical but some transitions are weak or organization is source-by-source rather than thematic",
167
+ "1": "Disjointed list of facts without narrative structure",
168
+ "0": "Incoherent structure"
169
+ }
170
+ }
171
+ ],
172
+ "expectedScoreWithout": 35,
173
+ "expectedScoreWith": 80
174
+ },
175
+ {
176
+ "id": "bench-med-02",
177
+ "difficulty": "medium",
178
+ "description": "Summarize a technical document preserving methodology and causal reasoning",
179
+ "input": "Summarize this research findings section, preserving the methodology and causal chain:\n\n---\nEffects of Sleep Deprivation on Cognitive Performance: A Meta-Analysis\n\nMethodology: We conducted a systematic review and meta-analysis of 147 studies (total N = 12,438 participants) published between 2010 and 2024, examining the relationship between sleep deprivation (defined as <6 hours of sleep per night for >=3 consecutive nights) and cognitive performance across five domains. Studies were identified through PubMed, PsycINFO, and Cochrane Library databases. Random-effects models were used to pool effect sizes, and heterogeneity was assessed using the I-squared statistic.\n\nFindings:\n\n1. Attention and Vigilance: The largest effect was observed in sustained attention tasks, with a pooled Cohen's d of -1.42 (95% CI: -1.58 to -1.26, I² = 34%). Psychomotor vigilance test (PVT) lapses increased by 287% after 3 nights of restricted sleep (<6h). Recovery required 2 nights of unrestricted sleep (mean 8.2 hours) to return to baseline.\n\n2. Working Memory: Moderate impairment was observed (d = -0.86, 95% CI: -1.01 to -0.71, I² = 42%). N-back task performance declined linearly: 2-back accuracy dropped from 89% to 67% after 5 nights of restriction. Notably, subjective awareness of impairment plateaued after night 3, while objective performance continued to decline — a phenomenon termed 'sleep debt denial.'\n\n3. Executive Function: Significant but variable effects (d = -0.73, 95% CI: -0.95 to -0.51, I² = 67%). High heterogeneity was driven by task type: planning tasks (Tower of London) showed d = -1.11, while cognitive flexibility tasks (Wisconsin Card Sort) showed only d = -0.38. This suggests that sleep deprivation disproportionately affects top-down, effortful processing.\n\n4. Emotional Regulation: Large effects on emotional processing (d = -1.18, 95% CI: -1.39 to -0.97, I² = 29%). Amygdala reactivity to negative stimuli increased by 60% (fMRI data from 23 studies), while prefrontal cortex regulatory activity decreased by 33%. This neural mechanism explains the consistent finding of increased irritability and impulsive decision-making.\n\n5. Long-term Memory Consolidation: Moderate effects (d = -0.91, 95% CI: -1.08 to -0.74, I² = 38%). Declarative memory (fact recall) was impaired by 23%, while procedural memory (motor skill learning) showed only 8% impairment. Sleep spindle density during NREM Stage 2, reduced by 41% under sleep restriction, correlated with declarative memory decline (r = 0.67, p < 0.001).\n\nDose-Response Relationship: Performance decline followed a nonlinear curve. The critical threshold was identified at 6.5 hours: participants sleeping 6-6.5 hours showed minimal impairment (d = -0.15), while those sleeping 5.5-6 hours showed moderate impairment (d = -0.62), and those below 5 hours showed severe impairment (d = -1.34). This nonlinear relationship has implications for public health guidelines.\n\nLimitations: 78% of studies used young adult samples (age 18-30), limiting generalizability. Only 12 studies examined chronic restriction beyond 7 days. Publication bias was detected (Egger's test p = 0.03), suggesting the true effects may be slightly smaller.\n---",
180
+ "rubric": [
181
+ {
182
+ "criterion": "Methodology Preservation",
183
+ "weight": 0.25,
184
+ "scoring": {
185
+ "5": "Retains: 147 studies, N=12,438, sleep deprivation definition (<6h for >=3 nights), 5 cognitive domains, meta-analytic approach",
186
+ "3": "Mentions it is a meta-analysis with study count but loses definition or scope details",
187
+ "1": "Vaguely mentions 'research' without methodology",
188
+ "0": "No methodology mentioned"
189
+ }
190
+ },
191
+ {
192
+ "criterion": "Finding Hierarchy",
193
+ "weight": 0.3,
194
+ "scoring": {
195
+ "5": "Preserves the ranked structure: attention most affected (d=-1.42), emotional regulation (d=-1.18), long-term memory (d=-0.91), working memory (d=-0.86), executive function (d=-0.73); includes at least 3 effect sizes",
196
+ "3": "Captures 3-4 domains with some effect sizes but loses ranking",
197
+ "1": "Lists domains without quantitative comparison",
198
+ "0": "Domains not differentiated"
199
+ }
200
+ },
201
+ {
202
+ "criterion": "Causal Mechanisms",
203
+ "weight": 0.25,
204
+ "scoring": {
205
+ "5": "Preserves key causal explanations: amygdala reactivity increase explaining emotional effects, sleep spindle reduction explaining memory decline, 'sleep debt denial' phenomenon, nonlinear dose-response with 6.5h threshold",
206
+ "3": "Captures 2 causal mechanisms",
207
+ "1": "Only reports effects without mechanisms",
208
+ "0": "No causal reasoning preserved"
209
+ }
210
+ },
211
+ {
212
+ "criterion": "Limitations Inclusion",
213
+ "weight": 0.2,
214
+ "scoring": {
215
+ "5": "Notes at least 2 of: young adult sample bias, limited chronic studies, publication bias",
216
+ "3": "Mentions limitations exist but lacks specifics",
217
+ "1": "No limitations mentioned",
218
+ "0": "Implies findings are unqualified/universal"
219
+ }
220
+ }
221
+ ],
222
+ "expectedScoreWithout": 35,
223
+ "expectedScoreWith": 82
224
+ },
225
+ {
226
+ "id": "bench-med-03",
227
+ "difficulty": "medium",
228
+ "description": "Multi-document summarization requiring claim reconciliation",
229
+ "input": "Synthesize the following three source excerpts into a unified summary:\n\nSource A (The Economist, January 2025):\n\"India's GDP growth slowed to 5.4% in Q3 2024, down from 6.7% in Q2, marking the weakest quarterly performance in two years. Manufacturing output contracted for the first time since 2020, falling 2.2% year-over-year. The Reserve Bank of India (RBI) maintained its repo rate at 6.5%, resisting calls for rate cuts despite the slowdown, citing persistent food inflation at 9.6%.\"\n\nSource B (IMF World Economic Outlook Update, January 2025):\n\"India remains the fastest-growing major economy, with projected full-year 2024 GDP growth of 6.5%, revised down from the October forecast of 7.0%. The medium-term outlook is robust, driven by demographic dividends, digitization, and infrastructure investment (National Infrastructure Pipeline: $1.4 trillion through 2027). However, the IMF flagged risks from elevated household debt (rising from 32% to 40% of GDP between 2019-2024) and a widening current account deficit.\"\n\nSource C (Reserve Bank of India Monetary Policy Statement, December 2024):\n\"The Indian economy continues to demonstrate resilience. While Q3 growth moderated due to temporary factors including monsoon disruptions and inventory adjustments in manufacturing, the RBI projects H2 FY2025 growth at 6.7-7.0%, supported by robust services sector performance and recovering rural demand. Food inflation, currently at 9.6%, is expected to moderate to 5.2% by Q1 FY2026 as supply-side pressures ease.\"\n\n---\nProvide a unified summary that reconciles these sources.",
230
+ "rubric": [
231
+ {
232
+ "criterion": "Claim Reconciliation",
233
+ "weight": 0.35,
234
+ "scoring": {
235
+ "5": "Identifies corroborated facts (5.4% Q3, 9.6% food inflation), contrasting framings (Economist: 'weakest', RBI: 'temporary'), and unique claims per source (IMF: household debt, RBI: H2 recovery projection)",
236
+ "3": "Reconciles some claims but misses key contradictions or agreements",
237
+ "1": "Presents sources separately without reconciliation",
238
+ "0": "Confuses or conflates source claims"
239
+ }
240
+ },
241
+ {
242
+ "criterion": "Source Attribution",
243
+ "weight": 0.25,
244
+ "scoring": {
245
+ "5": "Every claim correctly attributed to its source; framing differences noted with attribution (e.g., 'The Economist characterizes this as... while the RBI frames it as...')",
246
+ "3": "Most claims attributed but 1-2 unattributed or misattributed",
247
+ "1": "Generic 'sources say' without specific attribution",
248
+ "0": "No attribution or incorrect attribution"
249
+ }
250
+ },
251
+ {
252
+ "criterion": "Data Accuracy",
253
+ "weight": 0.2,
254
+ "scoring": {
255
+ "5": "Key numbers preserved accurately: 5.4% Q3, 6.5% full-year, -2.2% manufacturing, 6.5% repo rate, $1.4T infrastructure, 40% household debt",
256
+ "3": "4-5 key numbers correct",
257
+ "1": "1-2 numbers or numbers with errors",
258
+ "0": "Numbers fabricated or systematically wrong"
259
+ }
260
+ },
261
+ {
262
+ "criterion": "Thematic Organization",
263
+ "weight": 0.2,
264
+ "scoring": {
265
+ "5": "Organized by theme (current performance, outlook, risks) rather than by source; flows as a coherent narrative",
266
+ "3": "Partially thematic but some source-by-source structure remains",
267
+ "1": "Purely source-by-source with no integration",
268
+ "0": "Disorganized"
269
+ }
270
+ }
271
+ ],
272
+ "expectedScoreWithout": 30,
273
+ "expectedScoreWith": 78
274
+ },
275
+ {
276
+ "id": "bench-med-04",
277
+ "difficulty": "medium",
278
+ "description": "Summarize a narrative text preserving causal chain and key actors",
279
+ "input": "Summarize this case study, preserving the causal chain and key decisions:\n\n---\nThe Collapse of FTX: A Timeline of Decisions\n\nFTX, founded by Sam Bankman-Fried (SBF) in 2019, grew to become the world's third-largest cryptocurrency exchange with a peak valuation of $32 billion in January 2022. Its collapse in November 2022 represents one of the fastest destructions of value in financial history.\n\nThe seeds of collapse were planted in the organizational structure itself. Alameda Research, SBF's quantitative trading firm, maintained an undisclosed special relationship with FTX. Internal documents revealed that Alameda held a $65 billion line of credit from FTX — funded by customer deposits — with no collateral requirements, no independent risk oversight, and no board approval.\n\nThe trigger event occurred on November 2, 2022, when CoinDesk published a leaked Alameda balance sheet showing that $5.8 billion of its $14.6 billion in assets consisted of FTT tokens — FTX's own exchange token. This circular dependency meant that Alameda's solvency was contingent on the market price of a token controlled by its sister company.\n\nBinance CEO Changpeng Zhao (CZ) announced on November 6 that Binance would liquidate its $580 million FTT holdings, triggering a bank run. Within 72 hours, FTX customers attempted to withdraw $6 billion. FTX halted withdrawals on November 8 when it became clear that approximately $8 billion in customer funds had been transferred to Alameda and could not be returned.\n\nBinance signed a non-binding letter of intent to acquire FTX on November 8, but withdrew within 24 hours after due diligence revealed the scale of the shortfall. FTX filed for Chapter 11 bankruptcy on November 11, 2022. SBF was arrested in the Bahamas on December 12, 2022, and was convicted on all seven federal charges — including wire fraud and money laundering — on November 2, 2023. He was sentenced to 25 years in prison on March 28, 2024.\n\nThe bankruptcy proceedings, led by restructuring specialist John Ray III (who previously oversaw the Enron liquidation), recovered approximately $14.5 billion in assets — enough to repay all customers at the time of filing with interest. Ray described the corporate governance at FTX as the worst he had seen in his 40-year career.\n---",
280
+ "rubric": [
281
+ {
282
+ "criterion": "Causal Chain Preservation",
283
+ "weight": 0.35,
284
+ "scoring": {
285
+ "5": "Preserves the full chain: Alameda-FTX relationship (undisclosed credit line) → CoinDesk leak (FTT concentration) → CZ liquidation announcement → bank run → withdrawal halt → bankruptcy → conviction. Each link causally connected to the next",
286
+ "3": "Captures 4-5 links but some causal connections are implied rather than stated",
287
+ "1": "Lists events without causal linkage",
288
+ "0": "Causal chain broken or fabricated"
289
+ }
290
+ },
291
+ {
292
+ "criterion": "Key Figure Retention",
293
+ "weight": 0.2,
294
+ "scoring": {
295
+ "5": "Names SBF, CZ, and John Ray III with their roles; correctly attributes actions to each",
296
+ "3": "Names SBF and one other; actions mostly attributed correctly",
297
+ "1": "Only SBF named; others generic",
298
+ "0": "Key actors omitted or confused"
299
+ }
300
+ },
301
+ {
302
+ "criterion": "Critical Numbers",
303
+ "weight": 0.25,
304
+ "scoring": {
305
+ "5": "Retains: $32B valuation, $65B credit line, $5.8B FTT exposure, $6B withdrawal attempt, $8B shortfall, 25-year sentence, $14.5B recovery",
306
+ "3": "Retains 4-5 of these numbers",
307
+ "1": "Retains 1-2 numbers",
308
+ "0": "No numbers or numbers are wrong"
309
+ }
310
+ },
311
+ {
312
+ "criterion": "Compression Quality",
313
+ "weight": 0.2,
314
+ "scoring": {
315
+ "5": "Summary is 100-180 words, chronologically coherent, no filler",
316
+ "3": "Reasonable length but some chronological confusion or filler",
317
+ "1": "Too long (>250 words) or too short (<50 words) losing essential content",
318
+ "0": "Poor compression"
319
+ }
320
+ }
321
+ ],
322
+ "expectedScoreWithout": 40,
323
+ "expectedScoreWith": 85
324
+ },
325
+ {
326
+ "id": "bench-hard-01",
327
+ "difficulty": "hard",
328
+ "description": "Summarize a dense technical paper with complex argumentation and multiple competing frameworks",
329
+ "input": "Summarize the following research paper excerpt, preserving the methodological debate and the authors' contribution:\n\n---\nRethinking Attention: Sparse Mixture-of-Experts vs. Dense Transformers for Long-Context Understanding\n\nAbstract: We present SparseCtx, a novel architecture that replaces dense self-attention with a learned sparse routing mechanism for long-context tasks. On documents exceeding 100K tokens, SparseCtx achieves 94.2% of dense transformer accuracy while using only 18% of the compute (measured in FLOPs). We evaluate on five benchmarks: NarrativeQA, QuALITY, ScrollsQA, RULER-128K, and a new proprietary legal document analysis task.\n\nRelated Work and Methodological Debate: The efficient attention literature has bifurcated into two camps. Linear attention methods (Katharopoulos et al., 2020; Choromanski et al., 2021) replace softmax attention with kernel approximations, achieving O(n) complexity but sacrificing up to 15% accuracy on retrieval-intensive tasks (Tay et al., 2022). Sparse attention methods (Beltagy et al., 2020; Zaheer et al., 2020; Ainslie et al., 2023) maintain exact attention over selected token subsets, preserving accuracy but requiring careful selection heuristics that may miss critical long-range dependencies.\n\nOur work departs from both camps. Rather than approximating or subsetting attention, SparseCtx learns a routing function g(x) that maps each query token to k=32 expert attention heads (from a pool of 256), where each expert specializes in different dependency patterns: local context (window ≤512 tokens), medium-range (512-8K), long-range (8K-64K), and global (full context). The routing function is trained end-to-end with a load-balancing auxiliary loss (coefficient λ=0.01) to prevent expert collapse.\n\nKey Results:\n- NarrativeQA (F1): SparseCtx 71.8 vs. Dense 73.2 vs. Longformer 64.1 vs. Linear Attention 62.7\n- RULER-128K (accuracy): SparseCtx 89.4 vs. Dense 91.1 vs. Longformer 72.3 vs. Linear Attention 68.9\n- Inference latency at 128K tokens: SparseCtx 2.3s vs. Dense 12.8s vs. Longformer 4.1s\n- Training cost (GPU-hours for equivalent performance): SparseCtx 3,200 vs. Dense 18,400\n\nCritical Limitation: We observe a 'routing collapse' phenomenon in 3 of 20 training runs where the router degenerates to always selecting the same 4-5 experts regardless of input. While the load-balancing loss mitigates this, it does not eliminate it. We hypothesize this is an optimization landscape issue analogous to mode collapse in GANs and propose but do not validate a curriculum-based routing warmup strategy.\n---",
330
+ "rubric": [
331
+ {
332
+ "criterion": "Technical Accuracy",
333
+ "weight": 0.3,
334
+ "scoring": {
335
+ "5": "Correctly explains: sparse routing mechanism, the two existing camps (linear attention vs sparse attention) and their tradeoffs, how SparseCtx differs (learned routing to expert heads), and the routing collapse limitation",
336
+ "3": "Captures the main contribution but loses the methodological debate or specific mechanism details",
337
+ "1": "Vaguely describes it as an 'efficient attention method' without differentiation",
338
+ "0": "Technical content incorrect or fabricated"
339
+ }
340
+ },
341
+ {
342
+ "criterion": "Quantitative Results Fidelity",
343
+ "weight": 0.25,
344
+ "scoring": {
345
+ "5": "Preserves benchmark comparisons showing SparseCtx's position relative to dense and existing efficient methods; includes both accuracy and efficiency metrics (94.2% accuracy at 18% compute, latency comparison)",
346
+ "3": "Preserves some benchmark numbers but loses the comparative context",
347
+ "1": "Vague performance claims without numbers",
348
+ "0": "Numbers wrong or fabricated"
349
+ }
350
+ },
351
+ {
352
+ "criterion": "Contribution Clarity",
353
+ "weight": 0.25,
354
+ "scoring": {
355
+ "5": "Clearly articulates the paper's unique contribution: learned routing to specialized expert attention heads as a third approach beyond linear approximation and sparse subsetting, achieving near-dense accuracy at fraction of compute",
356
+ "3": "Describes what the paper does but doesn't clearly position it against alternatives",
357
+ "1": "Contribution unclear or lost in technical details",
358
+ "0": "Contribution misrepresented"
359
+ }
360
+ },
361
+ {
362
+ "criterion": "Limitation Honesty",
363
+ "weight": 0.2,
364
+ "scoring": {
365
+ "5": "Includes routing collapse issue (3/20 runs), notes it is not fully solved, mentions proposed but unvalidated mitigation",
366
+ "3": "Mentions limitations exist but lacks specifics",
367
+ "1": "No limitations mentioned, implying the approach is problem-free",
368
+ "0": "Ignores or obscures limitations"
369
+ }
370
+ }
371
+ ],
372
+ "expectedScoreWithout": 30,
373
+ "expectedScoreWith": 78
374
+ },
375
+ {
376
+ "id": "bench-hard-02",
377
+ "difficulty": "hard",
378
+ "description": "Summarize a long text with position bias trap — critical information buried in the middle and end",
379
+ "input": "Summarize the following report. Important: the most decision-relevant information is in sections 3 and 4, not the introduction.\n\n---\nQ4 2024 Market Analysis: Enterprise SaaS Sector\n\nSection 1 — Market Overview:\nThe global enterprise SaaS market reached $232 billion in 2024, growing 18% year-over-year. North America accounted for 52% of revenue ($120.6B), followed by Europe at 26% ($60.3B) and Asia-Pacific at 16% ($37.1B). Total venture capital investment in SaaS declined for the third consecutive year, falling to $41.2 billion from $52.8 billion in 2023 and $89.3 billion in 2022.\n\nSection 2 — Valuation Multiples:\nMedian EV/Revenue multiples for public SaaS companies compressed from 7.2x in January 2024 to 5.8x by December 2024. High-growth companies (>30% YoY revenue growth) maintained premiums at 12.4x, while sub-20% growth companies traded at 3.9x. The Rule of 40 (growth rate + profit margin) threshold separated premium from discount valuations: companies scoring >40 traded at 2.3x the multiple of those below 40.\n\nSection 3 — Churn Analysis (Critical for Strategic Planning):\nThe most significant trend of 2024 was the divergence in net revenue retention (NRR) between AI-augmented and traditional SaaS products. AI-augmented products achieved median NRR of 127%, compared to 108% for traditional products. More critically, gross churn for traditional products increased from 8.2% to 11.7% annually — the first time traditional SaaS churn exceeded 10% at the median.\n\nThe primary churn driver was identified through exit surveys (n=4,200 churned accounts): 43% cited replacement by AI-native alternatives, 28% cited budget consolidation, and 22% cited feature commoditization. The 43% AI-replacement figure represents a 3x increase from the 14% recorded in 2023 exit surveys.\n\nFor traditional SaaS companies, the data implies a 24-36 month window to integrate AI capabilities before churn reaches critical levels. Companies that launched AI features in 2024 saw churn reduce by an average of 3.1 percentage points within two quarters of launch.\n\nSection 4 — Emerging Risk: Usage-Based Pricing Erosion:\nA less-discussed but potentially more disruptive trend: AI-driven productivity gains are reducing per-seat and per-usage metrics that underpin SaaS pricing models. In customer support SaaS, AI automation reduced ticket volume by 40-60%, directly reducing usage-based revenue. CRM platforms report that AI-generated emails and call summaries reduced per-user activity metrics by 35%, triggering downgrades to lower pricing tiers.\n\nThis creates a paradox: the more effective a SaaS product's AI features, the more they cannibalize the product's own usage-based revenue. Five public SaaS companies revised their 2025 revenue guidance downward by 5-12% specifically citing AI-driven usage compression. The companies have not yet found a stable alternative pricing model, though value-based pricing (charging for outcomes rather than usage) is being piloted by 23% of surveyed companies.\n\nSection 5 — Outlook:\nThe consensus analyst forecast projects the SaaS market to reach $295 billion by 2026, but this projection assumes pricing models remain stable. If AI-driven usage compression spreads across categories, realized market size could be $250-265 billion, a 10-15% shortfall. The next 12 months will likely determine whether the industry can transition to new pricing paradigms.\n---",
380
+ "rubric": [
381
+ {
382
+ "criterion": "Position Bias Resistance",
383
+ "weight": 0.35,
384
+ "scoring": {
385
+ "5": "Summary prioritizes Sections 3-4 content (churn divergence AI vs traditional, 43% AI replacement, pricing erosion paradox) over the market overview numbers; demonstrates awareness that critical info was later in the document",
386
+ "3": "Covers Sections 3-4 but gives equal weight to the less critical Sections 1-2 market size numbers",
387
+ "1": "Dominated by Section 1-2 market overview; Sections 3-4 mentioned briefly or omitted",
388
+ "0": "Only summarizes the introduction"
389
+ }
390
+ },
391
+ {
392
+ "criterion": "Strategic Insight Extraction",
393
+ "weight": 0.3,
394
+ "scoring": {
395
+ "5": "Captures: NRR divergence (127% vs 108%), churn increase (8.2% to 11.7%), 24-36 month integration window, AI usage compression paradox, pricing model instability, and the 10-15% market size risk",
396
+ "3": "Captures 3-4 of these strategic insights",
397
+ "1": "Only surface-level observations without strategic implications",
398
+ "0": "Misses the strategic narrative"
399
+ }
400
+ },
401
+ {
402
+ "criterion": "Quantitative Precision",
403
+ "weight": 0.2,
404
+ "scoring": {
405
+ "5": "Retains key decision-relevant numbers: 43% AI replacement churn, 3.1pp churn reduction from AI launch, 40-60% ticket reduction, 35% activity metric reduction, 5-12% guidance revision, $250-265B adjusted forecast",
406
+ "3": "Retains 4-5 decision-relevant numbers",
407
+ "1": "Mostly retains overview numbers ($232B market size) rather than strategic ones",
408
+ "0": "Numbers absent or wrong"
409
+ }
410
+ },
411
+ {
412
+ "criterion": "Coherent Narrative",
413
+ "weight": 0.15,
414
+ "scoring": {
415
+ "5": "Tells a coherent story: AI is simultaneously driving growth (NRR boost) and threatening revenue (usage compression), creating urgency and paradox; the outlook depends on pricing model evolution",
416
+ "3": "Presents findings but doesn't connect them into a narrative",
417
+ "1": "Disconnected facts",
418
+ "0": "Incoherent"
419
+ }
420
+ }
421
+ ],
422
+ "expectedScoreWithout": 25,
423
+ "expectedScoreWith": 80
424
+ },
425
+ {
426
+ "id": "bench-hard-03",
427
+ "difficulty": "hard",
428
+ "description": "Summarize content with deliberate accuracy traps — similar numbers, easily confused entities, and hedged claims",
429
+ "input": "Summarize the following, being extremely careful with attribution and numbers:\n\n---\nGlobal AI Regulation: A Comparative Analysis of Three Frameworks\n\nThe European Union's AI Act, which entered into force on August 1, 2024, establishes a risk-based classification system with four tiers: unacceptable risk (banned), high risk (strict requirements), limited risk (transparency obligations), and minimal risk (no specific requirements). Penalties for non-compliance range from EUR 7.5 million or 1.5% of global turnover (for incorrect information) to EUR 35 million or 7% of global turnover (for banned practices). The Act requires high-risk AI systems to undergo conformity assessments, maintain technical documentation, and implement human oversight mechanisms. Implementation timelines are staggered: banned practices take effect February 2025, high-risk obligations August 2026, and general-purpose AI model requirements August 2025.\n\nChina's approach, codified through the Interim Measures for Generative AI Services (effective August 15, 2023) and the subsequent AI Safety Governance Framework (released September 2024), differs fundamentally from the EU model. Rather than risk tiers, China employs a sector-specific regulatory model where different ministries govern AI in their domains. The Cyberspace Administration of China (CAC) requires all generative AI services to undergo security assessments before public release and mandates that training data 'reflect core socialist values.' Penalties are comparatively modest: up to RMB 100,000 (approximately $14,000) per violation under the generative AI measures, though the Data Security Law enables penalties up to RMB 10 million ($1.4 million) for severe data-related violations. Unlike the EU Act, China's framework was in effect before the EU's and already governs over 180 registered generative AI services.\n\nThe United States has taken a decentralized approach. President Biden's Executive Order 14110 (October 30, 2023) established reporting requirements for foundation models trained using more than 10^26 FLOPs of compute and dual-use models with biological, cyber, or critical infrastructure capabilities. However, the EO was revoked by President Trump's Executive Order on January 20, 2025, which shifted policy toward 'removing barriers to American AI innovation.' As of early 2025, the US has no comprehensive federal AI legislation, though 45 states have introduced AI-related bills, with 17 passing into law. The most significant state action is the Colorado AI Act (effective February 2026), which mirrors some EU AI Act provisions for high-risk decision systems. The National Institute of Standards and Technology (NIST) AI Risk Management Framework (AI RMF 1.0, January 2023) serves as the primary voluntary guidance.\n\nComparative Assessment:\nEnforcement capacity varies dramatically. The EU has established a dedicated AI Office with 140 staff and a EUR 50 million annual budget. China's enforcement is distributed across existing regulators with no disclosed dedicated budget but demonstrated enforcement through the suspension of 14 AI services in 2024 for compliance failures. The US has no dedicated enforcement body for AI; oversight is fragmented across the FTC, FDA, DOT, and sector-specific regulators.\n\nInnovation impact assessments diverge as well. A Stanford HAI study estimated EU AI Act compliance costs at EUR 300,000-400,000 per high-risk system for SMEs, prompting 31% of surveyed European AI startups to consider relocating development to the US or UK. In contrast, a Tsinghua University survey found that 78% of Chinese AI companies reported that regulatory requirements 'improved their products' quality,' though critics note the survey's methodology did not account for survivorship bias (companies that exited the market due to regulation were not surveyed).\n---",
430
+ "rubric": [
431
+ {
432
+ "criterion": "Entity-Attribution Accuracy",
433
+ "weight": 0.35,
434
+ "scoring": {
435
+ "5": "Zero misattributions: EU penalties correctly stated (EUR 7.5M/1.5% to EUR 35M/7%), China penalties correctly stated (RMB 100K to RMB 10M), US EO 14110 correctly attributed to Biden and its revocation to Trump, Stanford HAI study not confused with Tsinghua study, Colorado AI Act not confused with EU AI Act",
436
+ "3": "1-2 minor attribution errors (e.g., wrong penalty range assigned to wrong framework)",
437
+ "1": "Multiple attribution errors or entities confused",
438
+ "0": "Systematic misattribution"
439
+ }
440
+ },
441
+ {
442
+ "criterion": "Framework Differentiation",
443
+ "weight": 0.25,
444
+ "scoring": {
445
+ "5": "Clearly distinguishes the three approaches: EU (risk-based tiers), China (sector-specific with values alignment), US (decentralized, currently no federal law); explains what makes each unique",
446
+ "3": "Distinguishes approaches but conflates some elements",
447
+ "1": "Treats frameworks as interchangeable or mischaracterizes their approach",
448
+ "0": "Fails to differentiate"
449
+ }
450
+ },
451
+ {
452
+ "criterion": "Nuance and Qualification Preservation",
453
+ "weight": 0.25,
454
+ "scoring": {
455
+ "5": "Preserves: staggered EU timelines (not just 'AI Act passed'), China framework existing before EU's, Biden EO revoked by Trump, Tsinghua survey's survivorship bias criticism, voluntary nature of NIST framework",
456
+ "3": "Captures 3-4 qualifications",
457
+ "1": "Presents findings without qualifications",
458
+ "0": "Oversimplifies to the point of inaccuracy"
459
+ }
460
+ },
461
+ {
462
+ "criterion": "Completeness at Appropriate Compression",
463
+ "weight": 0.15,
464
+ "scoring": {
465
+ "5": "Summary covers all three frameworks plus comparative assessment in 200-350 words; nothing critical omitted",
466
+ "3": "Covers all three frameworks but omits comparative section or is too long (>400 words)",
467
+ "1": "Omits one framework or the comparison entirely",
468
+ "0": "Covers only one framework"
469
+ }
470
+ }
471
+ ],
472
+ "expectedScoreWithout": 25,
473
+ "expectedScoreWith": 75
474
+ }
475
+ ]
476
+ }
@@ -0,0 +1,54 @@
1
+ {
2
+ "version": "0.0.1",
3
+ "timeout": 60,
4
+ "tasks": [
5
+ {
6
+ "id": "smoke-01",
7
+ "description": "Summarize a multi-paragraph argumentative article with quantitative data and competing claims",
8
+ "input": "Summarize the following article:\n\n---\nThe Future of Remote Work: A Data-Driven Analysis\n\nThe shift to remote work, accelerated by the COVID-19 pandemic, has become one of the most significant labor market transformations in decades. A 2024 Stanford study led by Nicholas Bloom, tracking 30,000 employees across 500 companies over 3 years, found that hybrid workers (3 days office, 2 days remote) showed a 13% productivity increase compared to full-time office workers, while fully remote workers showed only a 4% increase.\n\nHowever, the picture is not uniformly positive. Microsoft's 2024 Work Trend Index, surveying 31,000 workers across 31 countries, revealed that 68% of employees report not having enough uninterrupted focus time, and cross-team collaboration declined by 25% in fully remote settings. The report attributed this decline to reduced serendipitous interactions and weaker social ties between teams.\n\nCompensation dynamics have also shifted. According to Glassdoor's 2024 Salary Report, fully remote positions now command a 7-12% salary discount compared to equivalent hybrid roles in major metro areas, reversing the 2021-2022 trend where remote roles carried a premium. Economist Raj Chetty argues this discount reflects employers' perception of reduced oversight value rather than actual productivity differences.\n\nThe commercial real estate impact has been substantial. CBRE's Q3 2024 report shows US office vacancy rates at 19.6%, the highest since 1991, with Class B and C properties experiencing vacancy rates above 25%. JLL estimates that $1.2 trillion in commercial real estate value is at risk of repricing by 2026.\n\nMeanwhile, smaller cities have experienced population booms. The Brookings Institution reports that cities like Boise, Idaho and Bentonville, Arkansas saw population growth rates 3-4x the national average between 2020-2024, driven primarily by remote workers seeking lower cost of living.\n\nCritics of the remote-first model, including JPMorgan CEO Jamie Dimon, argue that in-person work is essential for mentorship, culture building, and career development. A 2024 Harvard Business Review study found that remote workers were 35% less likely to be promoted than their in-office counterparts, even when performance metrics were equivalent — a finding researchers attributed to proximity bias rather than performance differences.\n\nThe emerging consensus among labor economists is that hybrid models will dominate, with 60-70% of knowledge workers expected to work in hybrid arrangements by 2026. However, the optimal balance remains contested, with evidence supporting anywhere from 1 to 3 office days per week depending on role type, team structure, and organizational culture.\n---",
9
+ "rubric": [
10
+ {
11
+ "criterion": "Core Argument Capture",
12
+ "weight": 0.3,
13
+ "scoring": {
14
+ "5": "Summary captures the central thesis (hybrid model emerging as dominant), the productivity data supporting it, the counter-arguments (collaboration decline, promotion bias), and the contested optimal balance",
15
+ "3": "Captures the thesis and some supporting evidence but misses counter-arguments or nuance",
16
+ "1": "Only captures surface-level topic (remote work trends) without the argument structure",
17
+ "0": "Misidentifies or fabricates the central argument"
18
+ }
19
+ },
20
+ {
21
+ "criterion": "Quantitative Detail Retention",
22
+ "weight": 0.3,
23
+ "scoring": {
24
+ "5": "Retains key numbers: 13% productivity increase, 25% collaboration decline, 19.6% vacancy rate, 35% promotion gap, and at least 2 other critical statistics with proper attribution",
25
+ "3": "Retains 3-4 key statistics but drops some important ones or loses attribution",
26
+ "1": "Retains 1-2 numbers or replaces specific data with vague language (e.g., 'significant increase')",
27
+ "0": "No quantitative data preserved; all numbers replaced with qualitative language"
28
+ }
29
+ },
30
+ {
31
+ "criterion": "Attribution Accuracy",
32
+ "weight": 0.2,
33
+ "scoring": {
34
+ "5": "Correctly attributes findings to their sources (Stanford/Bloom, Microsoft, CBRE, HBR) and opinions to their holders (Dimon, Chetty); no misattribution",
35
+ "3": "Most attributions correct but 1-2 findings attributed to wrong source or unattributed",
36
+ "1": "Attributions mostly missing or generic ('studies show')",
37
+ "0": "Misattributes claims or fabricates sources"
38
+ }
39
+ },
40
+ {
41
+ "criterion": "Summary Structure and Compression",
42
+ "weight": 0.2,
43
+ "scoring": {
44
+ "5": "Well-structured summary at appropriate compression (5:1 to 10:1); uses formatting (bullets or paragraphs); leads with central finding; balanced coverage across all sections of the source",
45
+ "3": "Reasonable structure but unbalanced coverage (e.g., over-represents early paragraphs) or slightly too long/short",
46
+ "1": "Unstructured wall of text or extreme compression losing essential content",
47
+ "0": "No meaningful compression or disorganized output"
48
+ }
49
+ }
50
+ ],
51
+ "passThreshold": 60
52
+ }
53
+ ]
54
+ }