Compliance

Nick Shevelyov
CEO & managing partner
Published
Jan 8, 2026
If you feel like AI regulation is moving in two directions at once, you’re not imagining it.
On December 11, 2025, the White House issued an executive order aimed at reducing “state-law obstruction” of a national AI policy, calling for evaluations of state AI laws and a legislative recommendation for a uniform federal framework that could preempt conflicting state rules. (The White House)
At the same time, states are not standing still. In New York, the Responsible AI Safety and Education (RAISE) Act passed the state legislature in June 2025, and reporting indicates the governor has considered substantial rewrites—prompting pushback from supporters who want the bill signed without major changes. (The Verge)
This isn’t a politics blog. It’s an operations blog.
The lesson for founders and boards: don’t try to predict exactly how every AI law will land. Build an AI compliance posture that produces evidence.
The strategy that survives policy whiplash: “compliance as evidence”
When regulations are changing, you win by documenting three things:
What you are doing with AI (and what you are not doing)
Why you made those decisions (risk-based rationale)
How you monitor and respond when reality diverges from assumptions
That evidence becomes your defense in customer reviews, investor diligence, and—if needed—regulatory inquiries. (Consult counsel on jurisdiction-specific obligations.)
Minimum Viable AI Governance: a board-ready framework
Here is an AI governance framework that works for startups without turning into theater.
1) Maintain an AI Use Register (one page, always current)
For every AI-enabled feature or internal tool, record: owner (exec + product), data inputs (and whether they include sensitive categories), model/provider, human oversight, and monitoring plan.
2) Define “red lines” and “guardrails”
Red lines (board approval required): fully automated high-impact decisions without review; use of sensitive personal data without explicit governance; training/fine-tuning on customer content without clear permission.
Guardrails (allowed with controls): summarization, search, support drafting, controlled copilots on curated knowledge bases, assistive analytics with documented assumptions.
3) Vendor and model due diligence that asks AI-specific questions
Does the provider retain prompts/outputs? Is customer data used for training by default? What logs can we access? What controls protect connectors and admin roles? What happens during a provider incident?
4) AI incident response: treat “AI harm” like an incident class
Define triggers: sensitive data in outputs, material behavior change, customer harm reports, and regulatory inquiry. Have a playbook: isolate, preserve evidence, communicate, remediate.
5) Oversight: keep the board focused on decisions, not demos
Boards need answers: which AI use cases materially change risk, what evidence controls exist, and what you are not doing (and why).
30/60/90 days: an execution plan
In 30 days: publish the AI Use Register; identify sensitive data sources and block uncontrolled connectors; establish approval paths for new AI integrations.
In 60 days: run an “AI data exposure” tabletop; complete vendor AI due diligence; implement logging/monitoring for AI access paths.
In 90 days: produce a board memo (risks/controls/open decisions); operationalize a review cadence; test incident response end-to-end.
The calm, contrarian truth
Most “AI compliance” failures are not about laws. They’re about unmanaged scope: nobody owns the use cases, the data paths, or the exceptions.
If you own those three things, you can adapt as rules evolve.
If your board needs a defensible AI governance posture, vCSO.ai can build your AI Use Register, decision log, and 90-day control plan—aligned with your business model and risk tolerance.

Nick Shevelyov
CEO & managing partner




