How structured experimentation can transform government operations while maintaining public trust

State agencies, such as Departments of Motor Vehicles, Health, or Labor, are under mounting pressure to modernize services, improve efficiency, and manage risk as artificial intelligence (AI) transforms the public sector landscape. Yet, according to Code for America’s July 2025, Government AI Landscape Assessment, more than half of U.S. states are still at an “early” or “developing” stage in their AI journey, lacking the robust frameworks and hands-on experience needed for responsible adoption. For agency leaders, the imperative is clear: structured, safe experimentation is essential to move from aspiration to impact.

Key Reality Check: More than half of U.S. states are still at an “early” or “developing” stage in their AI journey, lacking the robust frameworks and hands-on experience needed for responsible adoption.

The Strategic Imperative for Safe Experimentation

Safe experimentation is the cornerstone of responsible AI adoption. As highlighted in a recently Harvard Business Review article, organizations that create structured environments for experimentation are more likely to innovate successfully while managing risk. Equally important is the idea that learning without doing is like trying to build muscle watching someone else exercise. On the other hand, state agencies have stakes for citizen safety, data privacy, and public trust that are exceptionally high, and experimentation without guardrails can be very risky.

Critical Insight:  AI fluency isn’t built through theory alone — it’s built through hands-on experimentation. The most fluent employees learn by engaging with AI in their real work.

Why it Matters for State Agencies

Most states are still building foundational AI capabilities. Without structured experimentation, agencies risk falling behind in service delivery and operational efficiency. Safe, hands-on learning accelerates both workforce readiness and organizational confidence.

Understanding Innovation Sandboxes: A Governance Framework

An innovation sandbox is more than a technical testbed, it’s a governance mechanism that enables rapid, safe, and transparent AI experimentation. Unlike traditional pilots, sandboxes provide:

    • secure, isolated environments for testing AI models and workflows without production system risks;
    • robust data governance and privacy controls to protect sensitive citizen information;
    • real-time monitoring and audit trails for accountability and compliance;
    • clear ethical guidelines and opt-in participation frameworks; and
    • integration capabilities with existing IT infrastructure and business processes.

The NIST AI Risk Management Framework offers a structured approach for sandbox governance, emphasizing four core functions of Govern, Map, Measure, and Manage.

Accelerated Implementation Timeline: Twelve to Eighteen Weeks to Value

State agencies can operationalize AI sandboxes in twelve to eighteen weeks through focused, cross-functional execution.

PhaseDurationKey Activities
Planning4 weeksStakeholder coalition building, risk assessment, policy drafting, hack-a-thon planning
Implementation4 weeksInfrastructure setup, security controls, integration with existing systems
Adoption & Training4-8 weeksUser onboarding, role-based training, live support, iterative feedback loops

Key Takeaway: With focused leadership and agile methods, agencies can move from concept to operational sandbox in under five months, without sacrificing rigor or safety.

Government Innovation in Action

Pennsylvania

The state’s generative AI pilot, launched under Governor Shapiro’s executive order, delivered measurable productivity gains and high employee satisfaction. The program emphasized human oversight, robust training, and a commitment to fairness and transparency.

New Jersey

The NJ AI Assistant provides a secure sandbox for state employees, with strong privacy protections and government-friendly terms of service. Its launch was paired with a comprehensive AI training program, and early results include a 50% increase in call center resolution rates.

Proven Results: New Jersey’s AI Assistant delivered a 50% increase in call center resolution rates, demonstrating the tangible impact of structured AI experimentation.

Virginia Tech

The Impact Makers Secure Research Environment, implemented for Virginia Tech, delivered a secure, cloud-based platform for researchers to work with sensitive data, collaborate, and deploy dedicated workspaces with built-in security and compliance. The solution was launched rapidly using agile methods and included features such as elastic infrastructure, automated monitoring, and HIPAA/NIH-compliant storage.

Launching with Innovation Hack-a-Thons

Tying sandbox availability for innovation hack-a-thons creates immediate momentum. Hack-a-thons surface real world use cases, foster cross-departmental collaboration, and provide rapid feedback on technical and governance challenges. This approach has been validated across multiple government implementations and academic institutions. 

Case in Point: The U.S. Congressional Hackathon united lawmakers, staff, and civic technologists to co-create tools that boost transparency, constituent engagement, and legislative workflows. Innovations included real-time bill comparison, AI-powered transcript search, and improved public access to budget data. These efforts have directly led to new tools like the centralized House calendar and enhanced social media tracking, helping modernize Congress’s digital infrastructure.

Toolkit for a High-Impact Launch

    • Secure executive sponsorship to signal commitment
    • Design for safety and transparency from day one
    • Plan a launch hack-a-thon to drive engagement and use case generation
    • Accelerate training with targeted, role-based programs
    • Iterate and scale based on early feedback and lessons learned

Building a Culture of Safe Experimentation

By leveraging innovation sandboxes, state agencies can foster a culture of safe, rapid experimentation, accelerating AI adoption while managing risk and building public trust. The evidence is clear. Those who experiment safely and learn fastest deliver the greatest value to their communities.

Read the Entire AI Adoption Series for State Agencies

For more practical insights, proven examples, and actionable strategies tailored for state agency leaders, check out the rest of our AI Adoption series below.

References