This is a preview of an UMMRO AI Safety Audit Report. Get Your Full Audit →
PREVIEW
AI Safety Audit Service
AI Safety Compliance
Audit Report
TechCorp AI Assistant — Multi-Model Security Assessment
Prepared For
TechCorp Inc.
Date
February 11, 2026
Report ID
UMR-2026-0211-TC
Author
Ahmed Adel Bakr Alderai
Section 01
Executive Summary
7.2
out of 10
B
Overall Risk Grade: Acceptable with Noted Deficiencies
Composite Compliance Score: 72%
10
Models Tested
5
Strategies / Model
72%
Avg. Safety Score
3
Critical Findings
UMMRO conducted a comprehensive AI safety compliance audit of the TechCorp AI Assistant deployment, evaluating 10 large language models across 5 distinct reframing strategies (50 total test scenarios). The overall safety posture is Acceptable (Grade B, 7.2/10), with strong performance from Claude 3.5 Sonnet and GPT-4o, but three models exhibited significant safety deficiencies requiring remediation.

Key Findings

Section 02
Compliance Scorecard
ModelProviderGradeScoreRefusal Rate
Claude 3.5 SonnetAnthropicA94%100%
GPT-4oOpenAIA91%96%
Gemini 2.5 ProGoogleB78%84%
Llama 3.3 70BMetaC52%56%
Gemma 2 27BGoogleF28%24%
Compliance Scorecard
Full model-by-model scores, refusal rates, latency benchmarks, and grade breakdowns for all 10 models.
Unlock Full Report
Section 03
Risk Matrix
ModelGuardrailRefusalSoft RefusalAdversarialQualityLatency
Claude 3.5 Sonnet98%96%94%92%90%95%
GPT-4o95%93%88%89%92%97%
Llama 3.3 70B54%52%42%38%68%82%
Gemma 2 27B22%28%18%20%42%86%
Risk Matrix Heat Map
Six-dimensional compliance heat map across all 10 models with color-coded risk assessment.
Unlock Full Report
Section 04
Model-by-Model Results

Detailed breakdown of each model's performance across all 5 reframing strategies, including per-strategy scores, refusal classifications, response previews, and latency measurements.

Model-by-Model Results
Detailed cards for each of 10 models with strategy-level scores, refusal breakdowns, and response analysis.
Unlock Full Report
Sections 05-07
Strategy Analysis, Compliance Mapping & Recommendations
5
Strategies Analyzed
4
Regulatory Frameworks
6
Recommendations
CRITICAL Remove Gemma 2 27B from Production
Model scored 28% with inverted refusal behavior...
HIGH Implement Role-Play Input Filtering
6 of 10 models vulnerable to persona-based prompts...
MEDIUM Standardize Safety in CI/CD Pipeline
Automated regression testing prevents safety drift...
Strategy Analysis, Compliance & Recommendations
Full reframing strategy effectiveness analysis, EU AI Act / NIST AI RMF / ISO 42001 / SOC 2 compliance mapping, and 6 prioritized remediation recommendations with effort estimates.
Unlock Full Report
Get Your Full AI Safety Audit
Know where your AI systems fail before regulators, customers, or adversaries find out.
$499
per audit • 48-hour delivery • branded PDF report
25+ AI models tested
17 reframing strategies
Risk grades A-F
Soft refusal detection
EU AI Act mapping
NIST AI RMF alignment
Actionable recommendations
Board-ready PDF
Get Your Audit Now

Free mini-audit available. No credit card required. Top 3 models tested free.