Skip to content

Commit d869f87

Browse files
authored
update agents.md h1 (#2273)
1 parent fc6e8c9 commit d869f87

File tree

1 file changed

+3
-26
lines changed

1 file changed

+3
-26
lines changed

examples/gpt-5/gpt-5-1-codex-max_prompting_guide.ipynb

Lines changed: 3 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -163,7 +163,7 @@
163163
"\n",
164164
"The Codex model family uses reasoning summaries to communicate user updates as it’s working. This can be in the form of one-liner headings (which updates the ephemeral text in Codex-CLI), or both heading and a short body. This is done by a separate model and therefore is **not promptable**, and we advise against adding any instructions to the prompt related to intermediate plans or messages to the user. We’ve improved these summaries for Codex-Max to be more communicative and provide more critical information about what’s happening and why; some of our users are updating their UX to promote these summaries more prominently in their UI, similar to how intermediate messages are displayed for GPT-5 series models.\n",
165165
"\n",
166-
"## [Agents.md](http://Agents.md) Usage\n",
166+
"## Using agents.md\n",
167167
"\n",
168168
"Codex-cli automatically enumerates these files and injects them into the conversation; the model has been trained to closely adhere to these instructions.\n",
169169
"\n",
@@ -185,7 +185,7 @@
185185
"\n",
186186
"# Compaction\n",
187187
"\n",
188-
"Compaction unlocks \"infinite\" context, where user conversations can persist for many turns without hitting context window limits or long context performance degradation, and agents can perform very long trajectories that exceed a typical context window for long-running, complex tasks. A weaker version of this was previously possible with ad-hoc scaffolding and conversation summarization, but our first-class implementation, available via the Responses API, is integrated with the model and is highly performant.\n",
188+
"Compaction unlocks significantly longer effective context windows, where user conversations can persist for many turns without hitting context window limits or long context performance degradation, and agents can perform very long trajectories that exceed a typical context window for long-running, complex tasks. A weaker version of this was previously possible with ad-hoc scaffolding and conversation summarization, but our first-class implementation, available via the Responses API, is integrated with the model and is highly performant.\n",
189189
"\n",
190190
"How it works:\n",
191191
"\n",
@@ -195,24 +195,7 @@
195195
" 2. The endpoint is ZDR compatible and will return an “encrypted\\_content” item that you can pass into future requests. \n",
196196
"3. For subsequent calls to the /responses endpoint, you can pass your updated, compacted list of conversation items (including the added compaction item). The model retains key prior state with fewer conversation tokens.\n",
197197
"\n",
198-
"**Endpoint:** POST /v1/responses/compact\n",
199-
"\n",
200-
"* Field: model \n",
201-
" * Type: string \n",
202-
" * Required: required \n",
203-
" * Notes: Use any Responses-compatible alias. \n",
204-
"* Field: input \n",
205-
" * Type: string or array of items \n",
206-
" * Required: optional \n",
207-
" * Notes: Provided messages (and optional tool/function items). Required unless `previous_response_id` is specified. \n",
208-
"* Field: previous\\_response\\_id \n",
209-
" * Type: string (response id) \n",
210-
" * Required: optional \n",
211-
" * Notes: Seed the run from an existing response; server hydrates its input items. \n",
212-
"* Field: Instructions \n",
213-
" * Type: string \n",
214-
" * Required: optional \n",
215-
" * Notes: Developer-style instructions forwarded to the compaction run.\n",
198+
"For endpoint details see our `/responses/compact` [docs](https://platform.openai.com/docs/api-reference/responses/compact).\n",
216199
"\n",
217200
"# Tools\n",
218201
"\n",
@@ -586,12 +569,6 @@
586569
"* Limit to 10k tokens. You can cheaply approximate this by computing `num_bytes/4`. \n",
587570
"* If you hit the truncation limit, you should use half of the budget for the beginning, half for the end, and truncate in the middle with `…3 tokens truncated…`\n"
588571
]
589-
},
590-
{
591-
"cell_type": "markdown",
592-
"id": "5f957bd4",
593-
"metadata": {},
594-
"source": []
595572
}
596573
],
597574
"metadata": {

0 commit comments

Comments
 (0)