meet emerging global AI regulations
Sixfold, the AI underwriting platform used by leading life, health, and P&C insurers, today announced
the release of its ‘Responsible AI Report 2025’, detailing how insurers can adopt, govern, and scale AI
in line with evolving global regulations.
With the EU AI Act and Colorado AI Act entering full enforcement in 2026, insurers will soon be
required to demonstrate fairness, oversight, and control across their AI systems.
“Compliance has shifted from principle to proof,” said Jane Tran, COO of Sixfold. “Regulators are no
longer asking whether insurers have AI policies. They’re asking companies to show how models are
governed, how bias is managed, and how oversight works in practice. Our report shows insurers how
to operationalise that for ease of compliance.”
The report defines five core pillars of Responsible AI that insurers can adopt or benchmark against:
- Bias prevention & fairness – fairness testing and controls that prevent reliance on protected
attributes - Model testing & validation – structured accuracy, robustness, and drift monitoring with
human oversight - AI governance & risk management – a governance programme aligned with global
requirements - Data security & integrity – SOC 2 Type II and HIPAA-aligned security, strict isolation, and data
stewardship - Explainability & transparency – clear traceability from model output to source evidence
“Insurers need systems they can defend to auditors and regulators,” said Alex Schmelkin, CEO of
Sixfold. “We built accountability into our daily operations so carriers can demonstrate continuous
assurance – not one-time compliance.”
Alex Schmelkin has spoken alongside regulators from the NAIC, Colorado Division of Insurance, and
Nebraska Department of Insurance, and serves on NAIC Committee H, focused on AI and emerging
technology governance.















