Navigating AI Model Safety Reviews: A Guide for Companies Partnering with Government Agencies
Overview
The landscape of artificial intelligence is rapidly evolving, and with it comes the need for robust safety oversight. Recently, major tech players including Google LLC, Microsoft Corp., and xAI have agreed to voluntarily share unreleased versions of their AI models with the U.S. Department of Commerce before public launch. This initiative is spearheaded by a new body: the Center for AI Standards and Innovation (CAISI), a division of the Commerce Department. CAISI will take the lead in testing these models to ensure they do not pose a threat to society. This guide is designed for AI companies—from startups to enterprises—looking to participate in this voluntary safety review process. It walks through the key steps, common pitfalls, and best practices to make your collaboration with CAISI as smooth as possible.

Prerequisites
Before engaging with CAISI for a model safety review, your organization must have several elements in place. These prerequisites are not formal requirements but essential for a productive partnership.
Understanding Regulatory Landscape
Familiarize your leadership and legal team with the current AI safety policies issued by the U.S. government, including executive orders related to AI. While the agreement is voluntary, being aligned with broader regulatory trends demonstrates good faith.
Internal Safety Protocols
Your organization should already maintain basic safety testing procedures. This includes red-teaming, bias audits, and performance benchmarks. CAISI’s review builds on your internal work, so having documented safety processes makes the collaboration more efficient.
Designated Points of Contact
Assign a dedicated team or individual responsible for liaising with CAISI. This person should be well-versed in both technical aspects of your model and regulatory compliance. They will manage the sharing of unreleased models and coordinate feedback loops.
Step-by-Step Instructions
Follow these steps to successfully participate in CAISI’s safety review process for your AI model. Note that exact procedures may vary as the initiative matures, but this framework is based on the announced agreement.
Step 1: Prepare Your Unreleased Model
Ensure your model is stable enough for external evaluation. This doesn't mean it must be fully polished, but it should be functionally complete enough to demonstrate its core capabilities and potential risks. Package the model in a secure, reproducible environment—preferably using containerization tools like Docker. Include version control details and a clear description of the training data, architecture, and intended use cases.
Step 2: Establish Communication with CAISI
Reach out to the Center for AI Standards and Innovation via its official channels (likely through the Commerce Department website). Express your interest in the voluntary safety review program. Provide a high-level overview of your model, its domain, and why you seek review. CAISI will assign a case manager to handle your submission.
Step 3: Share Model Access
Under the agreement, you will grant CAISI access to the unreleased model. This typically involves uploading the model to a secure government server or providing API access under a nondisclosure agreement. Follow CAISI’s instructions for data transfer—encrypt all communications and verify the receiving endpoint’s authenticity. The goal is to allow government safety experts to run their own tests without compromising your intellectual property.
Step 4: Submit Supporting Documentation
Alongside the model, provide detailed documentation covering:
- Model card: Standardized description of model performance, limitations, and fairness evaluations.
- Safety reports: Results from your internal red-teaming, bias checks, and stress tests.
- Deployment plan: How you intend to launch the model, including any content filters or usage restrictions.
- Risk mitigation strategy: Steps you’ve taken to prevent misuse, such as rate limiting, monitoring, and user authentication.
This documentation helps CAISI understand your model’s context and identify areas where deeper testing is needed.

Step 5: Participate in Testing Process
CAISI’s team will conduct a series of evaluations. You may be asked to assist by clarifying technical details, providing additional training data samples (without exposing sensitive data), or explaining unexpected behaviors. Be responsive and transparent—the process is collaborative. CAISI may run adversarial prompts, probe for hallucinations, assess bias across demographic groups, and check for vulnerabilities that could be exploited maliciously.
Step 6: Address Findings
After testing, CAISI will share their findings with you. This may include identified risks, performance gaps, or suggestions for improvement. You are not obligated to fix every issue, but to maintain the spirit of the agreement, you should seriously consider mitigating critical threats before launch. Work with your internal team to patch vulnerabilities, retrain on problematic areas, or implement guardrails. Submit revised versions if necessary.
Step 7: Receive Clearance for Release
Once CAISI is satisfied that your model does not pose a significant threat, they will issue a clearance statement. This does not mean government endorsement, but it signals that the model has passed a safety checkpoint. You can then proceed with public deployment. Keep documentation of the review process for future audits or voluntary disclosures.
Common Mistakes
Avoid these pitfalls to ensure a smooth safety review:
- Waiting until the last minute. Start the process early—ideally months before your planned release. CAISI’s testers need time to evaluate complex models, and last-minute fixes may not be feasible.
- Over-promising on safety. Don’t claim your model is completely safe. Be honest about known limitations. Exaggeration can erode trust during the review.
- Neglecting documentation. Sparse or unclear documentation forces CAISI to make assumptions, which can lead to extra rounds of questions and delays. Use established templates like model cards.
- Ignoring feedback. If CAISI finds a vulnerability but your organization dismisses it without justification, the process may stall. Even if you don’t agree, provide a rationale and suggest mitigations.
- Assuming clearance equals full immunity. Passing a safety review does not protect you from future regulatory actions or public scrutiny. Continue monitoring your model after release.
Summary
The voluntary agreement between Google, Microsoft, xAI, and the U.S. Commerce Department’s CAISI marks a new era of proactive AI safety. By following this guide—preparing your model, communicating with CAISI, sharing access, documenting thoroughly, and iterating on feedback—you can navigate the safety review process efficiently. Avoid common pitfalls like late submissions and poor documentation. Ultimately, this collaborative approach helps ensure that powerful AI technologies are launched responsibly, benefiting both companies and society at large.