agentic-team-templates 0.14.0 → 0.16.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/src/index.js +76 -11
- package/src/index.test.js +91 -1
- package/templates/educator/.cursorrules/accessibility.md +266 -0
- package/templates/educator/.cursorrules/assessment.md +215 -0
- package/templates/educator/.cursorrules/curriculum.md +286 -0
- package/templates/educator/.cursorrules/engagement.md +243 -0
- package/templates/educator/.cursorrules/instructional-design.md +235 -0
- package/templates/educator/.cursorrules/overview.md +91 -0
- package/templates/educator/.cursorrules/retention.md +235 -0
- package/templates/educator/CLAUDE.md +338 -0
|
@@ -0,0 +1,215 @@
|
|
|
1
|
+
# Assessment Design
|
|
2
|
+
|
|
3
|
+
Evidence-based approaches to measuring learning and guiding instruction.
|
|
4
|
+
|
|
5
|
+
## Assessment Types
|
|
6
|
+
|
|
7
|
+
### Formative vs. Summative
|
|
8
|
+
|
|
9
|
+
| Aspect | Formative | Summative |
|
|
10
|
+
|--------|-----------|-----------|
|
|
11
|
+
| Purpose | Guide instruction, identify gaps | Evaluate achievement |
|
|
12
|
+
| Timing | During learning | After learning |
|
|
13
|
+
| Stakes | Low (practice, not grading) | Higher (grading, certification) |
|
|
14
|
+
| Frequency | Continuous (every 10-15 min) | Periodic (end of unit/course) |
|
|
15
|
+
| Feedback | Immediate, specific, actionable | Evaluative, comparative |
|
|
16
|
+
| Analogy | GPS during the journey | Final destination photo |
|
|
17
|
+
|
|
18
|
+
### Formative Assessment Techniques
|
|
19
|
+
|
|
20
|
+
| Technique | Time | Description |
|
|
21
|
+
|-----------|------|-------------|
|
|
22
|
+
| Exit tickets | 2-3 min | Brief written response to a prompt at end of lesson |
|
|
23
|
+
| Think-pair-share | 3-5 min | Individual think → partner discuss → share with class |
|
|
24
|
+
| Muddiest point | 1-2 min | "What was most confusing today?" |
|
|
25
|
+
| One-minute paper | 1-2 min | "Summarize the most important idea from today" |
|
|
26
|
+
| Polling/clickers | 1 min | Real-time multiple choice with immediate data |
|
|
27
|
+
| Whiteboards | 2-3 min | All students display answers simultaneously |
|
|
28
|
+
| Fist-to-five | 30 sec | Self-assess confidence (0=lost, 5=mastered) |
|
|
29
|
+
| Misconception check | 2-3 min | Present common error; ask if correct and why |
|
|
30
|
+
|
|
31
|
+
### Summative Assessment Types
|
|
32
|
+
|
|
33
|
+
| Type | Best For | Bloom's Level |
|
|
34
|
+
|------|----------|---------------|
|
|
35
|
+
| Multiple choice | Recall, comprehension, discrimination | Remember–Analyze |
|
|
36
|
+
| Short answer | Recall, explanation, application | Remember–Apply |
|
|
37
|
+
| Essay | Analysis, evaluation, synthesis | Analyze–Create |
|
|
38
|
+
| Problem sets | Application, analysis | Apply–Analyze |
|
|
39
|
+
| Projects | Synthesis, creation, real-world transfer | Apply–Create |
|
|
40
|
+
| Portfolios | Growth over time, reflection, metacognition | All levels |
|
|
41
|
+
| Performances | Demonstration of skills in context | Apply–Create |
|
|
42
|
+
| Oral exams | Depth of understanding, reasoning process | Understand–Evaluate |
|
|
43
|
+
|
|
44
|
+
## Rubric Design
|
|
45
|
+
|
|
46
|
+
### Analytic vs. Holistic Rubrics
|
|
47
|
+
|
|
48
|
+
```markdown
|
|
49
|
+
Analytic Rubric: Scores each criterion separately
|
|
50
|
+
→ Use when: Detailed feedback needed, skills are independent
|
|
51
|
+
→ Example: Writing rubric (thesis, evidence, organization, grammar scored separately)
|
|
52
|
+
|
|
53
|
+
Holistic Rubric: Single overall score based on general impression
|
|
54
|
+
→ Use when: Quick scoring needed, overall quality matters more than components
|
|
55
|
+
→ Example: Art portfolio (overall artistic merit)
|
|
56
|
+
```
|
|
57
|
+
|
|
58
|
+
### Analytic Rubric Template
|
|
59
|
+
|
|
60
|
+
```markdown
|
|
61
|
+
| Criterion | Exemplary (4) | Proficient (3) | Developing (2) | Beginning (1) |
|
|
62
|
+
|-----------|---------------|-----------------|-----------------|----------------|
|
|
63
|
+
| [Criterion 1] | [Description of exemplary performance] | [Description of proficient performance] | [Description of developing performance] | [Description of beginning performance] |
|
|
64
|
+
| [Criterion 2] | [Description] | [Description] | [Description] | [Description] |
|
|
65
|
+
| [Criterion 3] | [Description] | [Description] | [Description] | [Description] |
|
|
66
|
+
```
|
|
67
|
+
|
|
68
|
+
### Rubric Best Practices
|
|
69
|
+
|
|
70
|
+
1. **Share rubrics before the assessment** — students should know the target
|
|
71
|
+
2. **Use descriptive language, not evaluative** — describe what the work looks like at each level
|
|
72
|
+
3. **Include examples** — anchor each level with sample work
|
|
73
|
+
4. **Limit criteria to 3-6** — more than 6 creates cognitive overload for assessors
|
|
74
|
+
5. **Use consistent scale** — same number of levels across criteria
|
|
75
|
+
6. **Involve learners** — co-create rubrics when appropriate
|
|
76
|
+
|
|
77
|
+
### Writing Rubric Descriptors
|
|
78
|
+
|
|
79
|
+
```markdown
|
|
80
|
+
❌ Vague: "Good use of evidence"
|
|
81
|
+
✅ Specific: "Integrates 3+ relevant sources with proper citations,
|
|
82
|
+
using evidence to directly support each claim in the argument"
|
|
83
|
+
|
|
84
|
+
❌ Comparative: "Better than average analysis"
|
|
85
|
+
✅ Absolute: "Identifies the underlying assumptions in the argument
|
|
86
|
+
and evaluates their validity using logical reasoning"
|
|
87
|
+
```
|
|
88
|
+
|
|
89
|
+
## Mastery-Based Progression
|
|
90
|
+
|
|
91
|
+
### Principle
|
|
92
|
+
|
|
93
|
+
Learners advance when they demonstrate mastery of objectives, not when a calendar date arrives.
|
|
94
|
+
|
|
95
|
+
### Mastery Learning Model (Bloom)
|
|
96
|
+
|
|
97
|
+
```
|
|
98
|
+
Instruction → Formative Assessment → Mastery?
|
|
99
|
+
│
|
|
100
|
+
┌─────────┴──────────┐
|
|
101
|
+
▼ ▼
|
|
102
|
+
Yes: Advance No: Corrective
|
|
103
|
+
to next unit instruction →
|
|
104
|
+
Re-assess →
|
|
105
|
+
Loop until mastery
|
|
106
|
+
```
|
|
107
|
+
|
|
108
|
+
### Mastery Criteria
|
|
109
|
+
|
|
110
|
+
```markdown
|
|
111
|
+
Define mastery BEFORE instruction:
|
|
112
|
+
- Criterion-referenced (not norm-referenced)
|
|
113
|
+
- Typically 80-90% accuracy on core objectives
|
|
114
|
+
- Must demonstrate on novel problems (not repeated items)
|
|
115
|
+
- Allow multiple attempts (learning from errors is the point)
|
|
116
|
+
```
|
|
117
|
+
|
|
118
|
+
### Mastery Gradebook
|
|
119
|
+
|
|
120
|
+
| Objective | Attempt 1 | Attempt 2 | Attempt 3 | Status |
|
|
121
|
+
|-----------|-----------|-----------|-----------|--------|
|
|
122
|
+
| Identify logical fallacies | 60% | 75% | 90% | Mastered |
|
|
123
|
+
| Construct valid arguments | 70% | 85% | — | Mastered |
|
|
124
|
+
| Evaluate source credibility | 55% | 65% | 70% | In Progress |
|
|
125
|
+
|
|
126
|
+
## Authentic Assessment
|
|
127
|
+
|
|
128
|
+
### Principle
|
|
129
|
+
|
|
130
|
+
Assessments should mirror real-world tasks that professionals or citizens actually perform.
|
|
131
|
+
|
|
132
|
+
### Authentic vs. Traditional
|
|
133
|
+
|
|
134
|
+
| Traditional | Authentic |
|
|
135
|
+
|-------------|-----------|
|
|
136
|
+
| Select a response | Construct a response |
|
|
137
|
+
| Contrived context | Real-world context |
|
|
138
|
+
| Recall/recognition | Application/transfer |
|
|
139
|
+
| Teacher as sole audience | Authentic audience |
|
|
140
|
+
| Single correct answer | Multiple valid approaches |
|
|
141
|
+
| One-time event | Ongoing process |
|
|
142
|
+
|
|
143
|
+
### Authentic Assessment Examples
|
|
144
|
+
|
|
145
|
+
| Subject | Traditional | Authentic |
|
|
146
|
+
|---------|-------------|-----------|
|
|
147
|
+
| Science | Multiple choice on lab procedures | Design and conduct an experiment |
|
|
148
|
+
| Writing | Grammar worksheet | Write a letter to the editor about a local issue |
|
|
149
|
+
| Math | Solve textbook problems | Create a budget for a community event |
|
|
150
|
+
| History | Memorize dates and events | Analyze primary sources to argue a historical thesis |
|
|
151
|
+
| Programming | Syntax quiz | Build a working application that solves a user need |
|
|
152
|
+
|
|
153
|
+
## Writing Effective Test Items
|
|
154
|
+
|
|
155
|
+
### Multiple Choice Best Practices
|
|
156
|
+
|
|
157
|
+
```markdown
|
|
158
|
+
✅ Good Item:
|
|
159
|
+
Stem: "Which of the following best explains why spaced practice
|
|
160
|
+
improves long-term retention?"
|
|
161
|
+
A) It reduces the total amount of study time needed
|
|
162
|
+
B) It forces repeated retrieval from long-term memory ← correct
|
|
163
|
+
C) It allows students to re-read material more times
|
|
164
|
+
D) It groups similar topics for efficient review
|
|
165
|
+
|
|
166
|
+
✅ Why it's good:
|
|
167
|
+
- Stem is a complete question
|
|
168
|
+
- All options are plausible
|
|
169
|
+
- No "all of the above" or "none of the above"
|
|
170
|
+
- Tests understanding, not just recall
|
|
171
|
+
- Options are parallel in structure and length
|
|
172
|
+
```
|
|
173
|
+
|
|
174
|
+
### Common Item-Writing Errors
|
|
175
|
+
|
|
176
|
+
| Error | Example | Fix |
|
|
177
|
+
|-------|---------|-----|
|
|
178
|
+
| Stem clue | "An unbiased sample is one that is NOT..." | Rephrase positively |
|
|
179
|
+
| Longest correct answer | Correct option has 3x the detail of distractors | Equalize length |
|
|
180
|
+
| Absolute terms | "always," "never" in distractors | Use qualified language |
|
|
181
|
+
| "All of the above" | Guessing 2 correct implies all correct | Remove; use select-all |
|
|
182
|
+
| Trivial content | "What year was X published?" | Test concepts, not trivia |
|
|
183
|
+
| Grammatical cue | "An ___" eliminates options starting with consonants | Check grammar |
|
|
184
|
+
|
|
185
|
+
## Feedback Design
|
|
186
|
+
|
|
187
|
+
### Effective Feedback Principles
|
|
188
|
+
|
|
189
|
+
1. **Timely** — as close to the performance as possible
|
|
190
|
+
2. **Specific** — points to exact elements, not just "good job"
|
|
191
|
+
3. **Actionable** — tells the learner what to do next
|
|
192
|
+
4. **Focused** — 2-3 points maximum per feedback instance
|
|
193
|
+
5. **Growth-oriented** — focuses on the work, not the person
|
|
194
|
+
|
|
195
|
+
### Feedback Framework
|
|
196
|
+
|
|
197
|
+
```markdown
|
|
198
|
+
1. What was done well (specific): "Your thesis statement clearly
|
|
199
|
+
states a debatable claim supported by your three main arguments."
|
|
200
|
+
|
|
201
|
+
2. What needs improvement (specific): "The second body paragraph
|
|
202
|
+
presents evidence but doesn't explain how it supports your thesis."
|
|
203
|
+
|
|
204
|
+
3. Next step (actionable): "Add 2-3 sentences after each piece of
|
|
205
|
+
evidence explaining the connection to your central argument."
|
|
206
|
+
```
|
|
207
|
+
|
|
208
|
+
### Feedback Timing
|
|
209
|
+
|
|
210
|
+
| When | Type | Purpose |
|
|
211
|
+
|------|------|---------|
|
|
212
|
+
| Immediate | Correct/incorrect with explanation | Factual knowledge, procedures |
|
|
213
|
+
| Delayed (hours) | Detailed written feedback | Complex tasks, essays, projects |
|
|
214
|
+
| Self-paced | Model answers for self-comparison | Building self-assessment skills |
|
|
215
|
+
| Peer | Structured peer review with rubric | Developing evaluative judgment |
|
|
@@ -0,0 +1,286 @@
|
|
|
1
|
+
# Curriculum Design
|
|
2
|
+
|
|
3
|
+
Frameworks for designing coherent, well-sequenced curricula that build deep understanding over time.
|
|
4
|
+
|
|
5
|
+
## Curriculum Mapping
|
|
6
|
+
|
|
7
|
+
### Scope and Sequence
|
|
8
|
+
|
|
9
|
+
```
|
|
10
|
+
Scope: WHAT content and skills are taught
|
|
11
|
+
Sequence: WHEN and in WHAT ORDER they are taught
|
|
12
|
+
|
|
13
|
+
Together: A complete map of the learning journey
|
|
14
|
+
```
|
|
15
|
+
|
|
16
|
+
### Curriculum Map Template
|
|
17
|
+
|
|
18
|
+
```markdown
|
|
19
|
+
| Unit | Duration | Standards | Essential Questions | Key Knowledge | Key Skills | Assessment |
|
|
20
|
+
|------|----------|-----------|--------------------| --------------|------------|------------|
|
|
21
|
+
| Unit 1 | 3 weeks | [Standard IDs] | "Why does X matter?" | [Concepts] | [Skills] | [Assessment type] |
|
|
22
|
+
| Unit 2 | 2 weeks | [Standard IDs] | "How does Y work?" | [Concepts] | [Skills] | [Assessment type] |
|
|
23
|
+
| Unit 3 | 4 weeks | [Standard IDs] | "What if Z changed?" | [Concepts] | [Skills] | [Assessment type] |
|
|
24
|
+
```
|
|
25
|
+
|
|
26
|
+
### Vertical Alignment
|
|
27
|
+
|
|
28
|
+
```
|
|
29
|
+
Grade/Level N-1: Foundational concepts (prerequisites)
|
|
30
|
+
↓
|
|
31
|
+
Grade/Level N: Current course (builds on N-1)
|
|
32
|
+
↓
|
|
33
|
+
Grade/Level N+1: Next course (builds on N)
|
|
34
|
+
|
|
35
|
+
Each level should:
|
|
36
|
+
- Know what came before (don't re-teach from scratch)
|
|
37
|
+
- Know what comes after (prepare learners for the next stage)
|
|
38
|
+
- Explicitly build on prior knowledge and skills
|
|
39
|
+
```
|
|
40
|
+
|
|
41
|
+
## Spiral Curriculum (Bruner)
|
|
42
|
+
|
|
43
|
+
### Principle
|
|
44
|
+
|
|
45
|
+
Revisit key concepts at increasing levels of complexity throughout the curriculum.
|
|
46
|
+
|
|
47
|
+
```
|
|
48
|
+
Complexity
|
|
49
|
+
▲
|
|
50
|
+
│ ╭─── Topic A (advanced application)
|
|
51
|
+
│ ╭────╯
|
|
52
|
+
│ ╭────╯ ╭─── Topic A (analysis & evaluation)
|
|
53
|
+
│ │ ╭───╯
|
|
54
|
+
│ │ ╭────╯ ╭─── Topic A (application)
|
|
55
|
+
│ │ │ ╭────╯
|
|
56
|
+
│ │ │ ╭────╯ ╭─── Topic A (introduction)
|
|
57
|
+
│ │ │ │ ╭────╯
|
|
58
|
+
├─┴────┴───┴────────┴──────────────────►
|
|
59
|
+
0 Unit 1 Unit 3 Unit 6 Unit 10 Time
|
|
60
|
+
```
|
|
61
|
+
|
|
62
|
+
### Spiral Design Principles
|
|
63
|
+
|
|
64
|
+
1. **Each revisit adds depth** — not just repetition, but new complexity
|
|
65
|
+
2. **Prior knowledge is activated** — explicitly connect to previous encounters
|
|
66
|
+
3. **Spacing is built in** — revisits are naturally spaced over time
|
|
67
|
+
4. **Multiple representations** — each revisit offers a different angle
|
|
68
|
+
5. **Assessment reflects growth** — later assessments expect higher Bloom's levels
|
|
69
|
+
|
|
70
|
+
### Spiral Curriculum Example
|
|
71
|
+
|
|
72
|
+
```markdown
|
|
73
|
+
Topic: Statistical Reasoning
|
|
74
|
+
|
|
75
|
+
Unit 2 (Remember/Understand):
|
|
76
|
+
→ Define mean, median, mode
|
|
77
|
+
→ Calculate measures of central tendency from a dataset
|
|
78
|
+
|
|
79
|
+
Unit 5 (Apply):
|
|
80
|
+
→ Choose the appropriate measure for different data distributions
|
|
81
|
+
→ Identify when mean is misleading (skewed data)
|
|
82
|
+
|
|
83
|
+
Unit 8 (Analyze):
|
|
84
|
+
→ Compare distributions using statistical measures
|
|
85
|
+
→ Analyze how sample size affects reliability
|
|
86
|
+
|
|
87
|
+
Unit 12 (Evaluate/Create):
|
|
88
|
+
→ Critique statistical claims in media reports
|
|
89
|
+
→ Design a study with appropriate statistical methods
|
|
90
|
+
```
|
|
91
|
+
|
|
92
|
+
## Prerequisite Mapping
|
|
93
|
+
|
|
94
|
+
### Dependency Graphs
|
|
95
|
+
|
|
96
|
+
```
|
|
97
|
+
[Advanced Topic D]
|
|
98
|
+
↑ ↑
|
|
99
|
+
[Topic B] [Topic C]
|
|
100
|
+
↑ ↑ ↑
|
|
101
|
+
[Topic A] [Topic A] [Topic A]
|
|
102
|
+
↑
|
|
103
|
+
[Prerequisite Knowledge]
|
|
104
|
+
```
|
|
105
|
+
|
|
106
|
+
### Creating Prerequisite Maps
|
|
107
|
+
|
|
108
|
+
```markdown
|
|
109
|
+
For each learning objective, ask:
|
|
110
|
+
1. What must the learner already KNOW to access this?
|
|
111
|
+
2. What must the learner already be able to DO?
|
|
112
|
+
3. What MISCONCEPTIONS might interfere?
|
|
113
|
+
|
|
114
|
+
Then:
|
|
115
|
+
1. Order objectives so prerequisites come first
|
|
116
|
+
2. Verify prerequisite skills with diagnostic assessment
|
|
117
|
+
3. Provide remediation paths for missing prerequisites
|
|
118
|
+
4. Make dependencies explicit to learners
|
|
119
|
+
```
|
|
120
|
+
|
|
121
|
+
### Prerequisite Verification
|
|
122
|
+
|
|
123
|
+
| Method | When | Purpose |
|
|
124
|
+
|--------|------|---------|
|
|
125
|
+
| Diagnostic pre-test | Start of unit/course | Identify missing prerequisites |
|
|
126
|
+
| Knowledge check | Start of each lesson | Verify yesterday's learning |
|
|
127
|
+
| Skills inventory | Start of course | Map individual readiness |
|
|
128
|
+
| Concept inventory | Start of unit | Identify misconceptions |
|
|
129
|
+
|
|
130
|
+
## Sequencing Principles
|
|
131
|
+
|
|
132
|
+
### Ordering Strategies
|
|
133
|
+
|
|
134
|
+
| Strategy | Description | When to Use |
|
|
135
|
+
|----------|-------------|-------------|
|
|
136
|
+
| Simple → Complex | Start with basic concepts, build toward complex ones | Skill-building, mathematics |
|
|
137
|
+
| Concrete → Abstract | Start with tangible examples, move to general principles | Conceptual understanding |
|
|
138
|
+
| Known → Unknown | Start with familiar context, extend to new territory | Connecting to prior knowledge |
|
|
139
|
+
| Chronological | Follow the historical or process timeline | History, procedures, narratives |
|
|
140
|
+
| Whole → Part → Whole | Overview first, then details, then synthesis | Systems thinking, complex topics |
|
|
141
|
+
| Problem-centered | Start with a problem, learn what's needed to solve it | Professional training, PBL |
|
|
142
|
+
|
|
143
|
+
### Chunking and Pacing
|
|
144
|
+
|
|
145
|
+
```markdown
|
|
146
|
+
Guidelines:
|
|
147
|
+
- 3-5 new concepts per session maximum (cognitive load)
|
|
148
|
+
- 15-20 minutes of new input before active processing
|
|
149
|
+
- Each chunk builds on the previous one
|
|
150
|
+
- Provide "landing points" where learners consolidate
|
|
151
|
+
|
|
152
|
+
Pacing Signals to Watch:
|
|
153
|
+
- Formative check shows < 60% comprehension → slow down, re-teach
|
|
154
|
+
- Formative check shows > 90% comprehension → accelerate or extend
|
|
155
|
+
- Mixed results → differentiate (some need support, some need extension)
|
|
156
|
+
```
|
|
157
|
+
|
|
158
|
+
## Continuous Improvement
|
|
159
|
+
|
|
160
|
+
### Data-Driven Curriculum Revision
|
|
161
|
+
|
|
162
|
+
```
|
|
163
|
+
Teach → Assess → Analyze → Adjust → Re-teach
|
|
164
|
+
↑ │
|
|
165
|
+
└──────────────────────────────────────┘
|
|
166
|
+
```
|
|
167
|
+
|
|
168
|
+
### Curriculum Review Cycle
|
|
169
|
+
|
|
170
|
+
| Frequency | Activity | Data Source |
|
|
171
|
+
|-----------|----------|-------------|
|
|
172
|
+
| Daily | Adjust lesson pacing | Formative assessment results |
|
|
173
|
+
| Weekly | Identify struggling objectives | Quiz/exit ticket analysis |
|
|
174
|
+
| Per Unit | Evaluate unit effectiveness | Summative assessment data |
|
|
175
|
+
| Semester | Review scope and sequence | Cumulative performance data |
|
|
176
|
+
| Annually | Major curriculum revision | Year-end data + student/teacher feedback |
|
|
177
|
+
|
|
178
|
+
### Questions for Curriculum Evaluation
|
|
179
|
+
|
|
180
|
+
```markdown
|
|
181
|
+
Effectiveness:
|
|
182
|
+
- Are learners meeting stated objectives? (assessment data)
|
|
183
|
+
- Which objectives have the lowest mastery rates? (identify gaps)
|
|
184
|
+
- Are there persistent misconceptions? (error analysis)
|
|
185
|
+
|
|
186
|
+
Alignment:
|
|
187
|
+
- Does each assessment measure its stated objective? (backward design check)
|
|
188
|
+
- Does each activity build toward an assessed objective? (activity audit)
|
|
189
|
+
- Are objectives appropriately sequenced? (prerequisite check)
|
|
190
|
+
|
|
191
|
+
Engagement:
|
|
192
|
+
- Where do learners disengage? (attendance, participation data)
|
|
193
|
+
- Which activities produce the deepest engagement? (observation, surveys)
|
|
194
|
+
- Are learners finding relevance? (student feedback)
|
|
195
|
+
|
|
196
|
+
Equity:
|
|
197
|
+
- Are achievement gaps present across groups? (disaggregated data)
|
|
198
|
+
- Are materials and examples inclusive? (content audit)
|
|
199
|
+
- Are all learners accessing support? (intervention data)
|
|
200
|
+
```
|
|
201
|
+
|
|
202
|
+
### Curriculum Documentation
|
|
203
|
+
|
|
204
|
+
```markdown
|
|
205
|
+
# Course: [Name]
|
|
206
|
+
|
|
207
|
+
## Course-Level Outcomes
|
|
208
|
+
By the end of this course, learners will be able to:
|
|
209
|
+
1. [Outcome aligned to program goals]
|
|
210
|
+
2. [Outcome aligned to program goals]
|
|
211
|
+
3. [Outcome aligned to program goals]
|
|
212
|
+
|
|
213
|
+
## Unit Map
|
|
214
|
+
|
|
215
|
+
### Unit 1: [Title] (Weeks 1-3)
|
|
216
|
+
- Objectives: [List]
|
|
217
|
+
- Prerequisites: [List or "None"]
|
|
218
|
+
- Key Vocabulary: [List]
|
|
219
|
+
- Assessments: Formative: [List], Summative: [Description]
|
|
220
|
+
- Spiral Connections: "Revisited in Unit 5 at Analyze level"
|
|
221
|
+
|
|
222
|
+
### Unit 2: [Title] (Weeks 4-5)
|
|
223
|
+
- Objectives: [List]
|
|
224
|
+
- Prerequisites: [Unit 1 objectives X and Y]
|
|
225
|
+
- ...
|
|
226
|
+
|
|
227
|
+
## Assessment Calendar
|
|
228
|
+
| Week | Formative | Summative |
|
|
229
|
+
|------|-----------|-----------|
|
|
230
|
+
| 1 | Daily exit tickets | — |
|
|
231
|
+
| 2 | Quiz 1 (Units 1a-1b) | — |
|
|
232
|
+
| 3 | Peer review | Unit 1 Project |
|
|
233
|
+
| ... | ... | ... |
|
|
234
|
+
|
|
235
|
+
## Revision Log
|
|
236
|
+
| Date | Change | Rationale | Evidence |
|
|
237
|
+
|------|--------|-----------|----------|
|
|
238
|
+
| [Date] | Moved Topic X before Topic Y | Students lacked prerequisite skills | Unit 2 pre-test data: 40% below threshold |
|
|
239
|
+
| [Date] | Added scaffolding to Unit 3 | High failure rate on summative | 35% of students scored below proficiency |
|
|
240
|
+
```
|
|
241
|
+
|
|
242
|
+
## Common Curriculum Pitfalls
|
|
243
|
+
|
|
244
|
+
### 1. Coverage Over Depth
|
|
245
|
+
|
|
246
|
+
```markdown
|
|
247
|
+
❌ "We need to cover 15 chapters this semester"
|
|
248
|
+
✅ "We need students to deeply understand 8 essential concepts"
|
|
249
|
+
|
|
250
|
+
Research (Schwartz et al.): Depth produces better transfer than breadth.
|
|
251
|
+
```
|
|
252
|
+
|
|
253
|
+
### 2. Activity-Driven Planning
|
|
254
|
+
|
|
255
|
+
```markdown
|
|
256
|
+
❌ "I found a great activity—let me build a lesson around it"
|
|
257
|
+
✅ "What's the objective? What assessment shows mastery? Now, what activity supports that?"
|
|
258
|
+
|
|
259
|
+
Activities serve objectives, not the other way around.
|
|
260
|
+
```
|
|
261
|
+
|
|
262
|
+
### 3. Teaching Topics Instead of Skills
|
|
263
|
+
|
|
264
|
+
```markdown
|
|
265
|
+
❌ "Week 4: World War II" (topic, not learning)
|
|
266
|
+
✅ "Week 4: Analyze how economic factors contributed to the rise of
|
|
267
|
+
authoritarian regimes in the 1930s" (skill + content)
|
|
268
|
+
```
|
|
269
|
+
|
|
270
|
+
### 4. Ignoring Prerequisite Gaps
|
|
271
|
+
|
|
272
|
+
```markdown
|
|
273
|
+
❌ Start unit → students fail → blame students
|
|
274
|
+
✅ Start unit → diagnostic pre-test → address gaps → proceed
|
|
275
|
+
|
|
276
|
+
You cannot build on a foundation that doesn't exist.
|
|
277
|
+
```
|
|
278
|
+
|
|
279
|
+
### 5. No Revision Process
|
|
280
|
+
|
|
281
|
+
```markdown
|
|
282
|
+
❌ Same curriculum year after year without data review
|
|
283
|
+
✅ Annual review cycle with student data driving changes
|
|
284
|
+
|
|
285
|
+
"The curriculum is a living document, not a monument."
|
|
286
|
+
```
|