terminal
howtonotcode.com
Codex app logo

Codex app

Ai Tool

A tool designed to assist developers with coding tasks using AI.

article 8 storys calendar_today First seen: 2026-02-03 update Last seen: 2026-02-14 open_in_new Website menu_book Wikipedia

Stories

Showing 1-8 of 8

OpenAI Skills + Shell for long‑running agents: patterns and pitfalls

OpenAI’s new Skills and Shell tooling make it easier to ship capability‑scoped, long‑running agents for real backend work, but early adopters report reliability gaps you should engineer around. OpenAI’s cookbook shows how to turn discrete capabilities into reusable Skills that your agent invokes via tool calls, enabling least‑privilege execution and clearer observability ([Skills in API](https://developers.openai.com/cookbook/examples/skills_in_api/)); paired with the “tool‑call render” pattern, this turns a chatty bot into a doer with predictable handoffs ([render pattern explainer](https://dev.to/programmingcentral/the-tool-call-render-pattern-turning-your-ai-from-a-chatty-bot-into-a-doer-4cb2)). For workloads that run minutes to hours, OpenAI’s guidance combines Shell, Skills, and compaction to manage state bloat, retry long steps, and keep transcripts affordable and debuggable ([Shell + Skills + Compaction tips](https://developers.openai.com/blog/skills-shell-tips/)). Plan for rough edges reported by developers: an embedding outage returned all‑zero vectors in text‑embedding‑3‑small, some Assistants API file uploads expired immediately, GPT‑5.2 extended‑thinking had very low tokens/sec for some, and Apps SDK toolInvocation status UI required a widget workaround ([embedding outage](https://community.openai.com/t/embedding-model-outage-text-embedding-3-small-api-ev3-model-name-with-all-0-values/1374079#post_10), [files expiring](https://community.openai.com/t/files-instantly-expiring-upon-upload/1366339#post_5), [slow generation](https://community.openai.com/t/gpt-5-2-extended-thinking-webchat-has-unworkably-slow-token-4-tps-generation/1373185?page=3#post_49), [toolInvocation UI bug](https://community.openai.com/t/bug-meta-openai-toolinvocation-invoking-and-meta-openai-toolinvocation-invoked-not-shown-unless-the-tool-registers-a-widget/1374087#post_1)).

calendar_today 2026-02-12
openai chatgpt assistants-api agents-sdk chatgpt-apps-sdk

OpenAI’s GPT-5.3-Codex rolls out to Copilot with faster, agentic workflows

OpenAI's GPT-5.3-Codex is a 25% faster, more agentic coding model built for long-running, tool-driven workflows and is now rolling out across Codex surfaces and GitHub Copilot with stronger cybersecurity guardrails. OpenAI positions the model for multi-step coding and broader "computer use" with SOTA benchmark results and notes early versions helped build and operate itself [Pulse 2.0](https://pulse2.com/openai-reveals-gpt-5-3-codex-a-faster-agentic-coding-model-built-for-long-running-work/)[^1] and [AI-360](https://www.ai-360.online/openai-launches-gpt-5-3-codex-extending-agentic-coding-and-real-time-steering/)[^2]. GitHub confirms GPT-5.3-Codex is GA in Copilot (Pro/Business/Enterprise) across VS Code, web, mobile, CLI, and the Coding Agent with an admin-enabled policy toggle and gradual rollout [GitHub Changelog](https://github.blog/changelog/2026-02-09-gpt-5-3-codex-is-now-generally-available-for-github-copilot/)[^3], while OpenAI channels have it now with API access "soon" and a new Trusted Access for Cyber pilot [Pulse 2.0](https://pulse2.com/openai-reveals-gpt-5-3-codex-a-faster-agentic-coding-model-built-for-long-running-work/)[^1] and [ITP.net](https://www.itp.net/ai-automation/openai-launches-gpt-5-3-codex-the-new-era-of-ai-powered-coding-and-beyond)[^4]. [^1]: Adds: Core capabilities, benchmark highlights, safety posture, availability across Codex app/CLI/IDE/web, and NVIDIA GB200 NVL72 infra. [^2]: Adds: Real-time steering in extended runs and cybersecurity classification/pilot context for enterprise adoption. [^3]: Adds: Concrete Copilot GA details, supported surfaces, plans, rollout, and admin policy enablement. [^4]: Adds: Additional context on broader professional task coverage and API timing.

calendar_today 2026-02-09
openai gpt-53-codex openai-codex-app github github-copilot

OpenAI recommends GPT-5.3-Codex as the default agentic coding model

OpenAI now recommends GPT-5.3-Codex as the default Codex model, signaling a step-up in agentic coding and reasoning for real-world engineering. The official Codex Models page highlights GPT-5.3-Codex as the most capable, with GPT-5.2-Codex as predecessor and a smaller GPT-5.1-Codex-mini option for cost-sensitive tasks [OpenAI Codex Models](https://developers.openai.com/codex/models/)[^1]. An anecdotal report describes spending $10,000 to automate research with Codex, indicating emerging large-scale usage patterns [Practitioner report](https://links.tldrnewsletter.com/J7poJAf substantial Codex-driven automation and spend.

calendar_today 2026-02-07
openai codex gpt-53-codex gpt-52-codex gpt-51-codex-mini

Codex 0.95–0.96 ship async compaction, rate-limit signals; MassGen adds Codex backend

OpenAI’s Codex app/server shipped 0.95–0.96 with v2 async thread compaction, websocket rate‑limit signaling, expanded skill loading/remote catalogs, shell parallelism, state‑DB correctness, telemetry, and Linux sandbox groundwork ([0.95.0](https://github.com/openai/codex/releases/tag/rust-v0.95.0)[^1], [0.96.0](https://github.com/openai/codex/releases/tag/rust-v0.96.0)[^2]). MassGen now offers a Codex backend with local/Docker modes to orchestrate multi‑agent workflows and MCP tooling ([MassGen v0.1.47](https://github.com/massgen/MassGen/releases/tag/v0.1.47)[^3]). Expect workflow differences vs IDEs—Codex is positioned as an agentic assistant, not a full IDE—and note a Windows PowerShell 5.1 ANSI‑encoding issue affecting Cyrillic output ([video](https://www.youtube.com/watch?v=ts7yQdfBW_U&pp=ygURQ3Vyc29yIElERSB1cGRhdGU%3D)[^4], [forum thread](https://community.openai.com/t/incorrect-cyrillic-rendering-in-codex-agent-on-windows-due-to-powershell-5-1-default-ansi-encoding/1356123#post_5)[^5]). [^1]: Release notes: skills loading and remote catalogs, macOS `codex app` CLI, shell parallelism, Git safety hardening, TUI improvements, Linux sandbox groundwork. [^2]: Release notes: `thread/compact` async RPC, websocket `codex.rate_limits` event, `unified_exec` enablement, state DB-first thread listing, telemetry. [^3]: MassGen adds a Codex backend (local/Docker), native tool architecture, and a quick start to try Codex workflows. [^4]: Explains Codex app’s agentic workflow vs IDEs like Cursor and how to use it effectively. [^5]: Documents Windows PowerShell 5.1 ANSI encoding causing Cyrillic rendering issues and workaround considerations.

calendar_today 2026-02-04
openai codex massgen cursor claude-code

Claude Code 2.1.x lands practical speedups and governed multi‑agent workflows

Anthropic pushed a rapid series of Claude Code 2.1 updates (v2.1.26–v2.1.31) that cut RAM on session resume, add page‑level PDF reads, support MCP servers without dynamic registration, enable PR‑based session bootstraps, and ship many reliability fixes [Reddit summary](https://www.reddit.com/r/ClaudeAI/comments/1qvgdc5/claude_code_v21262130_what_changed/)[^1] and [official v2.1.31 notes](https://github.com/anthropics/claude-code/releases/tag/v2.1.31)[^2]. Practitioners also highlight 2.1’s skill hot‑reload, lifecycle hooks, and forked sub‑agents as a foundation for governed, observable multi‑agent workflows—positioning Claude Code as a lightweight "agent OS" for real projects [deep dive](https://medium.com/@richardhightower/build-agent-skills-faster-with-claude-code-2-1-release-6d821d5b8179)[^3]. [^1]: Adds: community changelog for v2.1.26–30 covering performance, MCP, GitHub/PR workflows, and PDF handling. [^2]: Adds: official v2.1.31 fixes (PDF lockups, sandbox FS errors, streaming temperature override, tool routing prompts, provider labels) and hard limits (100 pages, 20MB). [^3]: Adds: perspective on skill hot‑reload, lifecycle hooks, and forked sub‑agents enabling governed multi‑agent patterns.

calendar_today 2026-02-04
claude-code anthropic mcp-model-context-protocol github slack

OpenAI ships Codex macOS app: multi-agent command center with git worktrees and skills

OpenAI introduced the macOS-only Codex app as a "command center" to run multiple coding agents in parallel, isolate work via git worktrees, and extend workflows with a new Skills system—plus a limited-time inclusion with ChatGPT Free/Go and doubled rate limits for paid plans ([OpenAI blog](https://openai.com/index/introducing-the-codex-app/?_bhlid=b040462c226c34eb9531cc536689e69b976397a7)[^1]). Developer docs confirm Apple Silicon support today, a Windows/Linux waitlist, and that API-key sign-in may limit features like cloud threads ([Codex app docs](https://developers.openai.com/codex/app/)[^2]). Reporting adds competitive context against Anthropic’s Code Cowork/Claude Code and notes model guidance (use GPT‑5.2‑Codex for coding) and multi-agent monitoring aimed at centralizing team workflows ([Fortune](https://fortune.com/2026/02/02/openai-launches-codex-app-to-bring-coding-models-to-more-users-openclaw-ai-agents/)[^3]). [^1]: Adds: official product details on multi-agent orchestration, git worktrees, Skills, and rate limit changes. [^2]: Adds: confirms macOS-only (Apple Silicon), Windows/Linux waitlist, and API-key limitations for cloud threads. [^3]: Adds: market context vs Anthropic, enterprise adoption, model recommendations, and multi-agent monitoring pitch.

calendar_today 2026-02-03
openai codex-app gpt-52-codex chatgpt anthropic