The acting leader of the U.S. Cybersecurity and Infrastructure Security Agency uploaded sensitive contracting documents marked “For Official Use Only” to ChatGPT, according to officials cited by Politico, triggering automated security alerts and prompting an internal review by the Department of Homeland Security. The acting director, Madhu Gottumukkala, had a rare exception to use the chatbot while most staff were barred, raising questions about oversight, risk management, and how federal agencies govern generative AI tools.
What Happened And Why It Matters For Federal AI Governance
Federal monitoring systems flagged the uploads as potential data loss events after the documents—unclassified but labeled for internal use—were pasted into a public version of the model. A CISA spokesperson told Politico the use was short-term and limited, and DHS officials moved to assess whether any harm resulted from the disclosures.

While “For Official Use Only” material does not carry a classified marking, it often contains procurement details, operational sensitivities, or partner information that agencies protect to reduce risk. Sharing such content with a public chatbot introduces potential exposure pathways: model providers may retain inputs for security and quality, and researchers have shown large language models can sometimes reproduce fragments of their training data under certain conditions. That combination creates a non-trivial risk of unintended disclosure.
The episode underscores a central tension in government security: the drive to leverage AI for productivity versus the duty to safeguard controlled information. It is especially notable given CISA’s role in setting the tone for cyber risk management across civilian agencies and critical infrastructure.
Policy Gaps And Federal AI Rules Under Scrutiny
Federal guidance on generative AI is maturing but uneven in implementation. The Office of Management and Budget has directed agencies to inventory use cases, apply safeguards, and restrict access to models that do not meet security requirements. NIST’s AI Risk Management Framework and Secure Software Development practices provide a baseline for mitigating risks like data leakage, supply chain exposure, and model misuse.
Agencies increasingly steer sensitive work toward enterprise-grade, FedRAMP-authorized AI services that commit to stronger data controls, including limits on retention and use for training. Public-facing chatbots, by contrast, remain off-limits for many government users absent explicit waivers. The fact that an exception was granted at the top of the nation’s civilian cyber agency will likely draw scrutiny from inspectors general and oversight committees.
A Government Accountability Office review has previously found that many agencies lack comprehensive processes to track AI use and enforce safeguards. This case will be a test of whether existing enterprise controls—like data loss prevention, traffic filtering, and exceptions governance—are working as designed when senior leaders seek flexibility.

A Sensitive Personnel Backdrop At CISA’s Leadership
Gottumukkala, appointed as CISA’s acting director after serving as South Dakota’s chief information officer under then-Governor Kristi Noem, has drawn attention beyond the ChatGPT uploads. He reportedly failed a counterintelligence polygraph that DHS later characterized as unsanctioned. He also suspended six long-serving staff members from access to classified information, a move that rattled parts of the agency. Those developments, combined with the AI incident, feed into a broader conversation about continuity, trust, and operational discipline at CISA during leadership transitions.
Precedents And The Real Risks Of LLM Data Exposure
The risks are not theoretical. High-profile mishaps have seen employees at major firms inadvertently paste proprietary code or internal notes into public chatbots, with subsequent clampdowns and enterprise AI rollouts to contain the fallout. Academic and industry teams, including researchers affiliated with Google, Stanford, and OpenAI, have demonstrated that language models can memorize and regurgitate rare or sensitive training snippets when prompted cleverly.
The human factor looms large. Verizon’s Data Breach Investigations Report has repeatedly found the majority of breaches involve the human element—misconfigurations, errors, or social engineering. Generative AI adds a new twist: the same tools that accelerate drafting and analysis can quietly expand the blast radius of a single copy-and-paste mistake. Here, even without classification markings, procurement specifics or internal processes can telegraph intent and capabilities to adversaries monitoring public models.
What Comes Next For DHS And CISA After The Incident
Expect a formal damage assessment, a review of exception-granting processes, and a renewed push for guardrails: enterprise-only access, logging, and pre-upload scanning for sensitive markings. Agencies are likely to double down on “safe AI” patterns—isolated government instances, strict retention controls, and redaction tools that strip controlled indicators before any external interaction.
Training and culture are just as critical. Clear, scenario-based guidance—what not to paste, which systems are authorized, and how to escalate gray areas—can reduce errors even among seasoned leaders. If this incident becomes a catalyst for better governance rather than a one-off embarrassment, it could ultimately strengthen federal posture on generative AI.
The bottom line: when the country’s chief civilian cyber agency wrestles with AI hygiene at the top, it is a warning shot for every public institution experimenting with these tools. Innovation cannot outrun policy, and exceptions must never outrun security.
