agentic-team-templates 0.14.0 → 0.16.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,215 @@
1
+ # Assessment Design
2
+
3
+ Evidence-based approaches to measuring learning and guiding instruction.
4
+
5
+ ## Assessment Types
6
+
7
+ ### Formative vs. Summative
8
+
9
+ | Aspect | Formative | Summative |
10
+ |--------|-----------|-----------|
11
+ | Purpose | Guide instruction, identify gaps | Evaluate achievement |
12
+ | Timing | During learning | After learning |
13
+ | Stakes | Low (practice, not grading) | Higher (grading, certification) |
14
+ | Frequency | Continuous (every 10-15 min) | Periodic (end of unit/course) |
15
+ | Feedback | Immediate, specific, actionable | Evaluative, comparative |
16
+ | Analogy | GPS during the journey | Final destination photo |
17
+
18
+ ### Formative Assessment Techniques
19
+
20
+ | Technique | Time | Description |
21
+ |-----------|------|-------------|
22
+ | Exit tickets | 2-3 min | Brief written response to a prompt at end of lesson |
23
+ | Think-pair-share | 3-5 min | Individual think → partner discuss → share with class |
24
+ | Muddiest point | 1-2 min | "What was most confusing today?" |
25
+ | One-minute paper | 1-2 min | "Summarize the most important idea from today" |
26
+ | Polling/clickers | 1 min | Real-time multiple choice with immediate data |
27
+ | Whiteboards | 2-3 min | All students display answers simultaneously |
28
+ | Fist-to-five | 30 sec | Self-assess confidence (0=lost, 5=mastered) |
29
+ | Misconception check | 2-3 min | Present common error; ask if correct and why |
30
+
31
+ ### Summative Assessment Types
32
+
33
+ | Type | Best For | Bloom's Level |
34
+ |------|----------|---------------|
35
+ | Multiple choice | Recall, comprehension, discrimination | Remember–Analyze |
36
+ | Short answer | Recall, explanation, application | Remember–Apply |
37
+ | Essay | Analysis, evaluation, synthesis | Analyze–Create |
38
+ | Problem sets | Application, analysis | Apply–Analyze |
39
+ | Projects | Synthesis, creation, real-world transfer | Apply–Create |
40
+ | Portfolios | Growth over time, reflection, metacognition | All levels |
41
+ | Performances | Demonstration of skills in context | Apply–Create |
42
+ | Oral exams | Depth of understanding, reasoning process | Understand–Evaluate |
43
+
44
+ ## Rubric Design
45
+
46
+ ### Analytic vs. Holistic Rubrics
47
+
48
+ ```markdown
49
+ Analytic Rubric: Scores each criterion separately
50
+ → Use when: Detailed feedback needed, skills are independent
51
+ → Example: Writing rubric (thesis, evidence, organization, grammar scored separately)
52
+
53
+ Holistic Rubric: Single overall score based on general impression
54
+ → Use when: Quick scoring needed, overall quality matters more than components
55
+ → Example: Art portfolio (overall artistic merit)
56
+ ```
57
+
58
+ ### Analytic Rubric Template
59
+
60
+ ```markdown
61
+ | Criterion | Exemplary (4) | Proficient (3) | Developing (2) | Beginning (1) |
62
+ |-----------|---------------|-----------------|-----------------|----------------|
63
+ | [Criterion 1] | [Description of exemplary performance] | [Description of proficient performance] | [Description of developing performance] | [Description of beginning performance] |
64
+ | [Criterion 2] | [Description] | [Description] | [Description] | [Description] |
65
+ | [Criterion 3] | [Description] | [Description] | [Description] | [Description] |
66
+ ```
67
+
68
+ ### Rubric Best Practices
69
+
70
+ 1. **Share rubrics before the assessment** — students should know the target
71
+ 2. **Use descriptive language, not evaluative** — describe what the work looks like at each level
72
+ 3. **Include examples** — anchor each level with sample work
73
+ 4. **Limit criteria to 3-6** — more than 6 creates cognitive overload for assessors
74
+ 5. **Use consistent scale** — same number of levels across criteria
75
+ 6. **Involve learners** — co-create rubrics when appropriate
76
+
77
+ ### Writing Rubric Descriptors
78
+
79
+ ```markdown
80
+ ❌ Vague: "Good use of evidence"
81
+ ✅ Specific: "Integrates 3+ relevant sources with proper citations,
82
+ using evidence to directly support each claim in the argument"
83
+
84
+ ❌ Comparative: "Better than average analysis"
85
+ ✅ Absolute: "Identifies the underlying assumptions in the argument
86
+ and evaluates their validity using logical reasoning"
87
+ ```
88
+
89
+ ## Mastery-Based Progression
90
+
91
+ ### Principle
92
+
93
+ Learners advance when they demonstrate mastery of objectives, not when a calendar date arrives.
94
+
95
+ ### Mastery Learning Model (Bloom)
96
+
97
+ ```
98
+ Instruction → Formative Assessment → Mastery?
99
+
100
+ ┌─────────┴──────────┐
101
+ ▼ ▼
102
+ Yes: Advance No: Corrective
103
+ to next unit instruction →
104
+ Re-assess →
105
+ Loop until mastery
106
+ ```
107
+
108
+ ### Mastery Criteria
109
+
110
+ ```markdown
111
+ Define mastery BEFORE instruction:
112
+ - Criterion-referenced (not norm-referenced)
113
+ - Typically 80-90% accuracy on core objectives
114
+ - Must demonstrate on novel problems (not repeated items)
115
+ - Allow multiple attempts (learning from errors is the point)
116
+ ```
117
+
118
+ ### Mastery Gradebook
119
+
120
+ | Objective | Attempt 1 | Attempt 2 | Attempt 3 | Status |
121
+ |-----------|-----------|-----------|-----------|--------|
122
+ | Identify logical fallacies | 60% | 75% | 90% | Mastered |
123
+ | Construct valid arguments | 70% | 85% | — | Mastered |
124
+ | Evaluate source credibility | 55% | 65% | 70% | In Progress |
125
+
126
+ ## Authentic Assessment
127
+
128
+ ### Principle
129
+
130
+ Assessments should mirror real-world tasks that professionals or citizens actually perform.
131
+
132
+ ### Authentic vs. Traditional
133
+
134
+ | Traditional | Authentic |
135
+ |-------------|-----------|
136
+ | Select a response | Construct a response |
137
+ | Contrived context | Real-world context |
138
+ | Recall/recognition | Application/transfer |
139
+ | Teacher as sole audience | Authentic audience |
140
+ | Single correct answer | Multiple valid approaches |
141
+ | One-time event | Ongoing process |
142
+
143
+ ### Authentic Assessment Examples
144
+
145
+ | Subject | Traditional | Authentic |
146
+ |---------|-------------|-----------|
147
+ | Science | Multiple choice on lab procedures | Design and conduct an experiment |
148
+ | Writing | Grammar worksheet | Write a letter to the editor about a local issue |
149
+ | Math | Solve textbook problems | Create a budget for a community event |
150
+ | History | Memorize dates and events | Analyze primary sources to argue a historical thesis |
151
+ | Programming | Syntax quiz | Build a working application that solves a user need |
152
+
153
+ ## Writing Effective Test Items
154
+
155
+ ### Multiple Choice Best Practices
156
+
157
+ ```markdown
158
+ ✅ Good Item:
159
+ Stem: "Which of the following best explains why spaced practice
160
+ improves long-term retention?"
161
+ A) It reduces the total amount of study time needed
162
+ B) It forces repeated retrieval from long-term memory ← correct
163
+ C) It allows students to re-read material more times
164
+ D) It groups similar topics for efficient review
165
+
166
+ ✅ Why it's good:
167
+ - Stem is a complete question
168
+ - All options are plausible
169
+ - No "all of the above" or "none of the above"
170
+ - Tests understanding, not just recall
171
+ - Options are parallel in structure and length
172
+ ```
173
+
174
+ ### Common Item-Writing Errors
175
+
176
+ | Error | Example | Fix |
177
+ |-------|---------|-----|
178
+ | Stem clue | "An unbiased sample is one that is NOT..." | Rephrase positively |
179
+ | Longest correct answer | Correct option has 3x the detail of distractors | Equalize length |
180
+ | Absolute terms | "always," "never" in distractors | Use qualified language |
181
+ | "All of the above" | Guessing 2 correct implies all correct | Remove; use select-all |
182
+ | Trivial content | "What year was X published?" | Test concepts, not trivia |
183
+ | Grammatical cue | "An ___" eliminates options starting with consonants | Check grammar |
184
+
185
+ ## Feedback Design
186
+
187
+ ### Effective Feedback Principles
188
+
189
+ 1. **Timely** — as close to the performance as possible
190
+ 2. **Specific** — points to exact elements, not just "good job"
191
+ 3. **Actionable** — tells the learner what to do next
192
+ 4. **Focused** — 2-3 points maximum per feedback instance
193
+ 5. **Growth-oriented** — focuses on the work, not the person
194
+
195
+ ### Feedback Framework
196
+
197
+ ```markdown
198
+ 1. What was done well (specific): "Your thesis statement clearly
199
+ states a debatable claim supported by your three main arguments."
200
+
201
+ 2. What needs improvement (specific): "The second body paragraph
202
+ presents evidence but doesn't explain how it supports your thesis."
203
+
204
+ 3. Next step (actionable): "Add 2-3 sentences after each piece of
205
+ evidence explaining the connection to your central argument."
206
+ ```
207
+
208
+ ### Feedback Timing
209
+
210
+ | When | Type | Purpose |
211
+ |------|------|---------|
212
+ | Immediate | Correct/incorrect with explanation | Factual knowledge, procedures |
213
+ | Delayed (hours) | Detailed written feedback | Complex tasks, essays, projects |
214
+ | Self-paced | Model answers for self-comparison | Building self-assessment skills |
215
+ | Peer | Structured peer review with rubric | Developing evaluative judgment |
@@ -0,0 +1,286 @@
1
+ # Curriculum Design
2
+
3
+ Frameworks for designing coherent, well-sequenced curricula that build deep understanding over time.
4
+
5
+ ## Curriculum Mapping
6
+
7
+ ### Scope and Sequence
8
+
9
+ ```
10
+ Scope: WHAT content and skills are taught
11
+ Sequence: WHEN and in WHAT ORDER they are taught
12
+
13
+ Together: A complete map of the learning journey
14
+ ```
15
+
16
+ ### Curriculum Map Template
17
+
18
+ ```markdown
19
+ | Unit | Duration | Standards | Essential Questions | Key Knowledge | Key Skills | Assessment |
20
+ |------|----------|-----------|--------------------| --------------|------------|------------|
21
+ | Unit 1 | 3 weeks | [Standard IDs] | "Why does X matter?" | [Concepts] | [Skills] | [Assessment type] |
22
+ | Unit 2 | 2 weeks | [Standard IDs] | "How does Y work?" | [Concepts] | [Skills] | [Assessment type] |
23
+ | Unit 3 | 4 weeks | [Standard IDs] | "What if Z changed?" | [Concepts] | [Skills] | [Assessment type] |
24
+ ```
25
+
26
+ ### Vertical Alignment
27
+
28
+ ```
29
+ Grade/Level N-1: Foundational concepts (prerequisites)
30
+
31
+ Grade/Level N: Current course (builds on N-1)
32
+
33
+ Grade/Level N+1: Next course (builds on N)
34
+
35
+ Each level should:
36
+ - Know what came before (don't re-teach from scratch)
37
+ - Know what comes after (prepare learners for the next stage)
38
+ - Explicitly build on prior knowledge and skills
39
+ ```
40
+
41
+ ## Spiral Curriculum (Bruner)
42
+
43
+ ### Principle
44
+
45
+ Revisit key concepts at increasing levels of complexity throughout the curriculum.
46
+
47
+ ```
48
+ Complexity
49
+
50
+ │ ╭─── Topic A (advanced application)
51
+ │ ╭────╯
52
+ │ ╭────╯ ╭─── Topic A (analysis & evaluation)
53
+ │ │ ╭───╯
54
+ │ │ ╭────╯ ╭─── Topic A (application)
55
+ │ │ │ ╭────╯
56
+ │ │ │ ╭────╯ ╭─── Topic A (introduction)
57
+ │ │ │ │ ╭────╯
58
+ ├─┴────┴───┴────────┴──────────────────►
59
+ 0 Unit 1 Unit 3 Unit 6 Unit 10 Time
60
+ ```
61
+
62
+ ### Spiral Design Principles
63
+
64
+ 1. **Each revisit adds depth** — not just repetition, but new complexity
65
+ 2. **Prior knowledge is activated** — explicitly connect to previous encounters
66
+ 3. **Spacing is built in** — revisits are naturally spaced over time
67
+ 4. **Multiple representations** — each revisit offers a different angle
68
+ 5. **Assessment reflects growth** — later assessments expect higher Bloom's levels
69
+
70
+ ### Spiral Curriculum Example
71
+
72
+ ```markdown
73
+ Topic: Statistical Reasoning
74
+
75
+ Unit 2 (Remember/Understand):
76
+ → Define mean, median, mode
77
+ → Calculate measures of central tendency from a dataset
78
+
79
+ Unit 5 (Apply):
80
+ → Choose the appropriate measure for different data distributions
81
+ → Identify when mean is misleading (skewed data)
82
+
83
+ Unit 8 (Analyze):
84
+ → Compare distributions using statistical measures
85
+ → Analyze how sample size affects reliability
86
+
87
+ Unit 12 (Evaluate/Create):
88
+ → Critique statistical claims in media reports
89
+ → Design a study with appropriate statistical methods
90
+ ```
91
+
92
+ ## Prerequisite Mapping
93
+
94
+ ### Dependency Graphs
95
+
96
+ ```
97
+ [Advanced Topic D]
98
+ ↑ ↑
99
+ [Topic B] [Topic C]
100
+ ↑ ↑ ↑
101
+ [Topic A] [Topic A] [Topic A]
102
+
103
+ [Prerequisite Knowledge]
104
+ ```
105
+
106
+ ### Creating Prerequisite Maps
107
+
108
+ ```markdown
109
+ For each learning objective, ask:
110
+ 1. What must the learner already KNOW to access this?
111
+ 2. What must the learner already be able to DO?
112
+ 3. What MISCONCEPTIONS might interfere?
113
+
114
+ Then:
115
+ 1. Order objectives so prerequisites come first
116
+ 2. Verify prerequisite skills with diagnostic assessment
117
+ 3. Provide remediation paths for missing prerequisites
118
+ 4. Make dependencies explicit to learners
119
+ ```
120
+
121
+ ### Prerequisite Verification
122
+
123
+ | Method | When | Purpose |
124
+ |--------|------|---------|
125
+ | Diagnostic pre-test | Start of unit/course | Identify missing prerequisites |
126
+ | Knowledge check | Start of each lesson | Verify yesterday's learning |
127
+ | Skills inventory | Start of course | Map individual readiness |
128
+ | Concept inventory | Start of unit | Identify misconceptions |
129
+
130
+ ## Sequencing Principles
131
+
132
+ ### Ordering Strategies
133
+
134
+ | Strategy | Description | When to Use |
135
+ |----------|-------------|-------------|
136
+ | Simple → Complex | Start with basic concepts, build toward complex ones | Skill-building, mathematics |
137
+ | Concrete → Abstract | Start with tangible examples, move to general principles | Conceptual understanding |
138
+ | Known → Unknown | Start with familiar context, extend to new territory | Connecting to prior knowledge |
139
+ | Chronological | Follow the historical or process timeline | History, procedures, narratives |
140
+ | Whole → Part → Whole | Overview first, then details, then synthesis | Systems thinking, complex topics |
141
+ | Problem-centered | Start with a problem, learn what's needed to solve it | Professional training, PBL |
142
+
143
+ ### Chunking and Pacing
144
+
145
+ ```markdown
146
+ Guidelines:
147
+ - 3-5 new concepts per session maximum (cognitive load)
148
+ - 15-20 minutes of new input before active processing
149
+ - Each chunk builds on the previous one
150
+ - Provide "landing points" where learners consolidate
151
+
152
+ Pacing Signals to Watch:
153
+ - Formative check shows < 60% comprehension → slow down, re-teach
154
+ - Formative check shows > 90% comprehension → accelerate or extend
155
+ - Mixed results → differentiate (some need support, some need extension)
156
+ ```
157
+
158
+ ## Continuous Improvement
159
+
160
+ ### Data-Driven Curriculum Revision
161
+
162
+ ```
163
+ Teach → Assess → Analyze → Adjust → Re-teach
164
+ ↑ │
165
+ └──────────────────────────────────────┘
166
+ ```
167
+
168
+ ### Curriculum Review Cycle
169
+
170
+ | Frequency | Activity | Data Source |
171
+ |-----------|----------|-------------|
172
+ | Daily | Adjust lesson pacing | Formative assessment results |
173
+ | Weekly | Identify struggling objectives | Quiz/exit ticket analysis |
174
+ | Per Unit | Evaluate unit effectiveness | Summative assessment data |
175
+ | Semester | Review scope and sequence | Cumulative performance data |
176
+ | Annually | Major curriculum revision | Year-end data + student/teacher feedback |
177
+
178
+ ### Questions for Curriculum Evaluation
179
+
180
+ ```markdown
181
+ Effectiveness:
182
+ - Are learners meeting stated objectives? (assessment data)
183
+ - Which objectives have the lowest mastery rates? (identify gaps)
184
+ - Are there persistent misconceptions? (error analysis)
185
+
186
+ Alignment:
187
+ - Does each assessment measure its stated objective? (backward design check)
188
+ - Does each activity build toward an assessed objective? (activity audit)
189
+ - Are objectives appropriately sequenced? (prerequisite check)
190
+
191
+ Engagement:
192
+ - Where do learners disengage? (attendance, participation data)
193
+ - Which activities produce the deepest engagement? (observation, surveys)
194
+ - Are learners finding relevance? (student feedback)
195
+
196
+ Equity:
197
+ - Are achievement gaps present across groups? (disaggregated data)
198
+ - Are materials and examples inclusive? (content audit)
199
+ - Are all learners accessing support? (intervention data)
200
+ ```
201
+
202
+ ### Curriculum Documentation
203
+
204
+ ```markdown
205
+ # Course: [Name]
206
+
207
+ ## Course-Level Outcomes
208
+ By the end of this course, learners will be able to:
209
+ 1. [Outcome aligned to program goals]
210
+ 2. [Outcome aligned to program goals]
211
+ 3. [Outcome aligned to program goals]
212
+
213
+ ## Unit Map
214
+
215
+ ### Unit 1: [Title] (Weeks 1-3)
216
+ - Objectives: [List]
217
+ - Prerequisites: [List or "None"]
218
+ - Key Vocabulary: [List]
219
+ - Assessments: Formative: [List], Summative: [Description]
220
+ - Spiral Connections: "Revisited in Unit 5 at Analyze level"
221
+
222
+ ### Unit 2: [Title] (Weeks 4-5)
223
+ - Objectives: [List]
224
+ - Prerequisites: [Unit 1 objectives X and Y]
225
+ - ...
226
+
227
+ ## Assessment Calendar
228
+ | Week | Formative | Summative |
229
+ |------|-----------|-----------|
230
+ | 1 | Daily exit tickets | — |
231
+ | 2 | Quiz 1 (Units 1a-1b) | — |
232
+ | 3 | Peer review | Unit 1 Project |
233
+ | ... | ... | ... |
234
+
235
+ ## Revision Log
236
+ | Date | Change | Rationale | Evidence |
237
+ |------|--------|-----------|----------|
238
+ | [Date] | Moved Topic X before Topic Y | Students lacked prerequisite skills | Unit 2 pre-test data: 40% below threshold |
239
+ | [Date] | Added scaffolding to Unit 3 | High failure rate on summative | 35% of students scored below proficiency |
240
+ ```
241
+
242
+ ## Common Curriculum Pitfalls
243
+
244
+ ### 1. Coverage Over Depth
245
+
246
+ ```markdown
247
+ ❌ "We need to cover 15 chapters this semester"
248
+ ✅ "We need students to deeply understand 8 essential concepts"
249
+
250
+ Research (Schwartz et al.): Depth produces better transfer than breadth.
251
+ ```
252
+
253
+ ### 2. Activity-Driven Planning
254
+
255
+ ```markdown
256
+ ❌ "I found a great activity—let me build a lesson around it"
257
+ ✅ "What's the objective? What assessment shows mastery? Now, what activity supports that?"
258
+
259
+ Activities serve objectives, not the other way around.
260
+ ```
261
+
262
+ ### 3. Teaching Topics Instead of Skills
263
+
264
+ ```markdown
265
+ ❌ "Week 4: World War II" (topic, not learning)
266
+ ✅ "Week 4: Analyze how economic factors contributed to the rise of
267
+ authoritarian regimes in the 1930s" (skill + content)
268
+ ```
269
+
270
+ ### 4. Ignoring Prerequisite Gaps
271
+
272
+ ```markdown
273
+ ❌ Start unit → students fail → blame students
274
+ ✅ Start unit → diagnostic pre-test → address gaps → proceed
275
+
276
+ You cannot build on a foundation that doesn't exist.
277
+ ```
278
+
279
+ ### 5. No Revision Process
280
+
281
+ ```markdown
282
+ ❌ Same curriculum year after year without data review
283
+ ✅ Annual review cycle with student data driving changes
284
+
285
+ "The curriculum is a living document, not a monument."
286
+ ```