Краткий вывод
- Best for high-complexity reasoning, planning, and code analysis workflows.
- Uses OpenAI-compatible format: POST /v1/chat/completions for low-friction SDK migration.
- Supports stream=true SSE output for IDE copilots and real-time assistants.
Ключевые параметры
- model | string | обяз. | gpt-5-4-official | - | Model ID for this page (for example gpt-5-4-official).
- messages | object[] | обяз. | - | - | Conversation messages in chronological order with system/user/assistant roles.
- max_tokens | integer | опц. | - | >=1 | Maximum output tokens (model default applies when omitted).
- stream | boolean | опц. | false | - | Whether to enable SSE streaming output.
- temperature | number | опц. | 1 | 0-2 | Sampling temperature controlling randomness.
- top_p | number | опц. | 1 | 0-1 | Nucleus sampling threshold; avoid aggressively tuning with temperature together.
- stop | string | string[] | опц. | - | - | Stop sequence(s), up to 4 entries.
- Authorization | HTTP Header | обяз. | - | - | Bearer auth: Authorization: Bearer <YOUR_API_KEY>.
Частые ошибки
- 400 invalid_request_error: триггер=Missing required fields or invalid field types in payload.; исправление=Validate model, messages, and parameter types.; повтор=Retry only after fixing payload.
- 401 authentication_error: триггер=Missing/invalid auth header or invalid API key.; исправление=Verify Authorization header format and key validity.; повтор=Retry after auth is fixed.
- 429 rate_limit_error: триггер=Request rate, concurrency, or current quota hits upstream rate limiting.; исправление=Apply exponential backoff first, then review request rate, concurrency, and quota usage.; повтор=Use 1s/2s/4s backoff with jitter; if it persists, reduce submission pressure.

