waypoint-codex 1.0.3 → 1.0.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "waypoint-codex",
3
- "version": "1.0.3",
3
+ "version": "1.0.4",
4
4
  "description": "Make Codex better by default with stronger planning, code quality, reviews, tracking, and repo guidance.",
5
5
  "license": "MIT",
6
6
  "type": "module",
@@ -105,6 +105,40 @@ The bottleneck here is not token thrift inside Codex. The bottleneck is giving G
105
105
  Curate relevance aggressively. Compress relevance only by excluding things that truly do not matter.
106
106
  Do **not** compress relevant context just because it is long.
107
107
 
108
+ ## Core Rule: Make The Request Standalone
109
+
110
+ Write the handoff so GPT-5.4-Pro can produce a strong answer in a brand new session with no prior context and no follow-up.
111
+
112
+ Do not rely on:
113
+
114
+ - prior chat history
115
+ - "as discussed above"
116
+ - local repo knowledge
117
+ - implicit shared assumptions
118
+ - GPT-5.4-Pro asking clarifying questions before it can reason well
119
+
120
+ The request should contain the full problem, the full context, the real objective, and the expected output shape.
121
+
122
+ ## Core Rule: Ask Outcome-Level Questions
123
+
124
+ Frame the request around the real goal, source of truth, or decision to be made.
125
+
126
+ Prefer:
127
+
128
+ - "Given the attached context, propose the best architecture for this system"
129
+ - "Evaluate the current implementation against the attached north-star document and identify all important mismatches"
130
+ - "Given the current product, constraints, and code, recommend the best migration strategy"
131
+
132
+ Avoid narrow, history-led framing unless the task is intentionally narrow.
133
+
134
+ Avoid prompts like:
135
+
136
+ - "We recently changed X, can you check if this looks right?"
137
+ - "Did we fix issue Y?"
138
+ - "Can you look at this follow-up from the last pass?"
139
+
140
+ Prior findings, recent changes, and local debates can be included as background, but they should not become the main frame unless that is truly the task.
141
+
108
142
  ## Workflow
109
143
 
110
144
  ### 1. Justify AGI-Help
@@ -133,6 +167,7 @@ Collect the context it would need to reason well, such as:
133
167
  - what has already been tried
134
168
  - what is blocked, unclear, risky, or contentious
135
169
  - what a successful answer would help us decide or do next
170
+ - what source of truth, target state, desired behavior, or decision standard should govern the answer
136
171
 
137
172
  This is not a fixed checklist. Include whatever materially changes the quality of the answer.
138
173
 
@@ -154,6 +189,8 @@ Examples of relevant files:
154
189
  - failing or partial implementations
155
190
  - screenshots or exported artifacts when available through the current tool surface
156
191
  - strategy docs, briefs, drafts, notes, or prior outputs that define the problem
192
+ - source-of-truth documents such as specs, north-star docs, acceptance criteria, or decision memos
193
+ - surrounding files that materially affect the verdict even if they were not changed recently
157
194
 
158
195
  Preserve relative structure inside `files/` when it helps orientation.
159
196
 
@@ -172,6 +209,8 @@ Keep this concise but useful.
172
209
 
173
210
  Write the prompt as if briefing a world-class expert who has zero implicit context.
174
211
 
212
+ The prompt should ask for a complete answer to the actual problem, not a reaction to the latest local narrative.
213
+
175
214
  The prompt should usually include:
176
215
 
177
216
  1. **Role / framing**
@@ -189,7 +228,7 @@ The prompt should usually include:
189
228
  7. **Attached materials**
190
229
  - tell it that files are attached and should be read before answering
191
230
  8. **Specific request**
192
- - the concrete question or task
231
+ - the concrete task framed at the outcome/spec/decision level
193
232
  9. **Desired output shape**
194
233
  - exactly how the answer should be structured
195
234
 
@@ -212,6 +251,14 @@ Ask for something concrete, such as:
212
251
  - a better strategy or positioning approach
213
252
  - a decision memo with tradeoffs and risks
214
253
 
254
+ When the task is evaluative, ask for a complete evaluation against the governing standard, not confirmation of recent fixes.
255
+
256
+ Examples:
257
+
258
+ - "Does the current system satisfy the attached target architecture? If not, identify all important mismatches with evidence."
259
+ - "Given the attached implementation and requirements, what is the best recommendation and why?"
260
+ - "Using the attached code, docs, and constraints, produce a complete architecture proposal for the new system."
261
+
215
262
  ### Specify The Output Format
216
263
 
217
264
  Tell GPT-5.4-Pro how to respond.
@@ -226,6 +273,17 @@ Good example shapes:
226
273
 
227
274
  Explicitly instruct it to review the attached files before answering.
228
275
 
276
+ ### Do Not Depend On Follow-Up
277
+
278
+ Assume Mark wants one strong answer, not a clarification loop.
279
+
280
+ So:
281
+
282
+ - front-load the necessary context
283
+ - state the objective precisely
284
+ - include the governing constraints and source of truth
285
+ - ask for the full answer in one pass
286
+
229
287
  ## Final Handoff To Mark
230
288
 
231
289
  When the bundle is ready, report: