Security researchers at Phantom Labs discovered a command injection vulnerability in OpenAI's Codex that could have allowed attackers to steal GitHub authentication tokens from developers using ChatGPT's coding assistant. The flaw worked by tricking Codex into executing malicious commands that would exfiltrate sensitive credentials, potentially giving attackers access to private repositories and development environments. BeyondTrust, Phantom Labs' parent company, reported the vulnerability has since been patched by OpenAI.
This vulnerability highlights a fundamental tension in AI coding assistants: the more capable they become at executing code and interacting with development environments, the larger their attack surface grows. Codex isn't just generating code suggestions—it's actively interfacing with developer toolchains, including version control systems where authentication tokens are critical assets. The fact that a command injection could compromise these tokens shows how traditional security vulnerabilities are finding new vectors through AI systems.
While the original reporting focused on the technical mechanics of the exploit, the broader implications extend beyond this single bug. Every major coding assistant—from GitHub Copilot to Claude's code execution—now sits between developers and their most sensitive development resources. The vulnerability was apparently discovered through responsible disclosure, but it raises questions about how thoroughly these AI systems are being audited for injection attacks and other security flaws that could weaponize their growing capabilities against the very developers they're designed to help.
