How to Assess Risks in AI Development Before Committing Budget

Content Writer

Dave Quinn
Head of Software Engineering

Reviewer

Dave Quinn
Head of Software Engineering

Table of Contents


Use a structured risk assessment framework aligned with the NIST AI Risk Management Framework when evaluating AI projects requiring more than 6 months delivery or involving regulated data. Reactive assessment works for single use case prototypes under 3 months with internal data only. Structured assessment prevents the majority of AI project failures by identifying data quality, compliance, and capability gaps before budget commitment.

Key Takeaways
  • Structured assessment reduces AI project failure rates from 80% to under 30%. Teams using frameworks like NIST AI RMF identify budget, compliance, and capability gaps 4 to 6 weeks before reactive teams discover the same issues.
  • Data quality assessment must happen first, consuming 60 to 80% of initial evaluation time. According to RAND Corporation research, poor data is among the leading causes of AI project failure.
  • European SMBs face higher regulatory complexity than US counterparts. GDPR and the AI Act create compliance obligations that must be assessed before technical feasibility.

Why This Framework Matters

CTOs, Heads of Engineering, and VP Product roles at European SMBs face AI investment decisions with incomplete information. Research from RAND Corporation shows that over 80% of AI projects fail before reaching production, with failure rates significantly higher than traditional software projects. The difference between successful and failed projects often comes down to risk assessment timing, not technical capability.

Budget decisions made without structured risk assessment lead to three common outcomes. Teams discover data quality issues 4 months into development, forcing scope reduction or abandonment. Regulatory compliance requirements surface during vendor security reviews, adding 3 to 6 months of unexpected work. Capability gaps appear when prototypes reach production, requiring emergency hiring or external partnerships.

The stakes increase for SMBs operating with constrained budgets. A failed AI project does not just waste money. It consumes 8 to 12 months of engineering time, damages stakeholder confidence, and hands competitive advantage to rivals who execute successfully. The AI Act introduces requirements that shift risk profiles for high-risk AI systems, making proactive assessment more valuable than reactive responses.


Step 1: Audit Data Quality and Availability

What it is: A systematic review of data completeness, accuracy, consistency, and accessibility across all systems that will feed the AI project. This audit identifies gaps, duplicates, format inconsistencies, and access restrictions before any development work begins.

Why it matters for budget decisions: Poor data quality causes a disproportionate share of AI project abandonments. Data problems that surface 4 months into development force scope changes or complete abandonment. Budget estimates that assume clean data consistently underestimate actual effort by 40 to 60%.

How to do it

  • Document every data source the AI project will use, including CRM systems, databases, spreadsheets, third-party APIs, and manual entry processes
  • Calculate completeness rates for each critical field (aim for 95% or higher for fields the model depends on)
  • Identify duplicate records across systems (more than 10% duplication indicates poor governance)
  • Test data format consistency (dates, currencies, addresses must follow standardised patterns)
  • Map access controls and permissions to confirm engineering teams can reach production data
  • Estimate data remediation effort in weeks, not days

Red flags to watch for

  • Customer or transaction data lives in 4 or more disconnected systems with no integration
  • More than 15% of critical fields are empty or contain placeholder values
  • No documented data dictionary exists, and business teams define fields differently than technical teams
  • Data access requires approval processes that take longer than 5 business days

Decision threshold: If data remediation effort exceeds 25% of total project budget, address data quality as a separate foundational project before committing AI budget.


Step 2: Map Regulatory and Compliance Requirements

What it is: Identifying every regulation, standard, and compliance obligation that applies to your AI project based on industry sector, data types, and geographic markets. For European SMBs, this includes GDPR, the AI Act, and sector-specific frameworks like financial services regulations or healthcare standards.

Why it matters for budget decisions: Compliance obligations discovered mid-project add 3 to 6 months of unexpected work. The AI Act classifies certain AI systems as high-risk, triggering mandatory risk management, documentation, and testing requirements. European SMBs operating across borders face fragmented regulatory requirements that increase complexity compared to single-market operations.

How to do it

  • Determine if your AI system processes personal data (triggers GDPR requirements)
  • Check if the AI Act classifies your use case as high-risk (employment decisions, credit scoring, essential services, biometric identification)
  • Identify sector-specific regulations that apply (financial services, healthcare, insurance each have distinct AI governance requirements)
  • Document cross-border data transfer requirements if serving customers in multiple EU markets
  • Review customer contracts for vendor security and compliance obligations
  • Budget 2 to 4 months for compliance assessment and documentation if operating in regulated sectors

Red flags to watch for

  • Your AI will process health data, financial data, or employment decisions without a Data Protection Officer
  • Customer contracts require ISO 27001 or SOC 2 certification but your organisation lacks these
  • The AI system makes automated decisions affecting individuals without human oversight capability
  • Cross-border data flows involve countries outside the EU without adequacy decisions

Decision threshold: If compliance requirements add more than 30% to technical development timeline, factor this into ROI calculations. Projects with payback periods extending beyond 18 months often fail to maintain stakeholder support.


Step 3: Assess Internal Capability Gaps

What it is: Evaluating whether your current team has the specific skills, experience, and capacity to build, deploy, and maintain production AI systems. This assessment covers ML engineering, data engineering, MLOps, and domain expertise.

Why it matters for budget decisions: According to McKinsey’s State of AI research, skills gaps remain among the top obstacles for organisations scaling AI. Hiring timelines for senior ML engineers average 6 to 9 months in European markets. Skills gaps discovered after budget approval force emergency hiring at premium rates or rushed external partnerships that deliver suboptimal outcomes.

How to do it

  • Count engineers with production ML deployment experience (not just coursework or prototypes)
  • Verify team capacity by reviewing current project commitments (AI projects fail when treated as side work)
  • Assess MLOps capability separately from ML capability (deploying models requires different skills than building them)
  • Calculate the minimum viable team size (production AI typically requires 2 to 3 specialists: ML engineer, data engineer, MLOps)
  • Compare internal hiring timelines against project deadlines
  • Evaluate build versus partner tradeoffs based on urgency and long-term capability goals

Red flags to watch for

  • Zero engineers on the team have deployed ML models to production environments handling real user traffic
  • Your data scientist spends more than 50% of time on infrastructure tasks instead of model development
  • No formal training budget exists for ML frameworks that evolve quarterly
  • The project timeline assumes AI work fits alongside existing full-time responsibilities

Decision threshold: If hiring timelines exceed 6 months or if zero production ML experience exists internally, external AI engineering partnerships deliver capability faster than building from scratch. Teams under 50 employees rarely sustain dedicated AI specialists economically.


Step 4: Model Infrastructure and Operational Requirements

What it is: Calculating the full infrastructure requirements for development, training, deployment, and ongoing operations. AI workloads require GPU compute, storage, networking, and monitoring that exceed traditional application requirements by 3 to 10 times.

Why it matters for budget decisions: Infrastructure requirements that surprise teams 3 to 4 months into projects force scope reduction or abandonment. Cloud spending for AI can spike unexpectedly without proper monitoring and controls. Production ML systems require uptime, auto-scaling, and redundancy that many SMBs underestimate.

How to do it

  • Calculate GPU compute requirements for model training (estimate hours required and multiply by cloud provider GPU instance rates)
  • Model inference requirements based on expected request volume and latency targets
  • Budget for development, staging, and production environments separately
  • Include monitoring, logging, and observability tools in operational planning
  • Factor in data storage requirements that grow as training datasets expand
  • Add 30% contingency for overruns during initial production months

Red flags to watch for

  • Plans assume AI runs on existing application infrastructure without separate capacity
  • No monitoring or alerting configured on cloud accounts
  • Model inference latency targets conflict with instance type allocations
  • Auto-scaling policies exist without spending caps

Decision threshold: If monthly operational requirements for production AI exceed 15% of expected revenue impact, revisit the business case. Infrastructure requirements that never stabilise indicate architectural problems requiring expert intervention.


Step 5: Define Success Metrics and Failure Thresholds

What it is: Establishing concrete, measurable outcomes that define project success and explicit thresholds that trigger project re-evaluation or cancellation. Success metrics must exist before development starts, not after deployment.

Why it matters for budget decisions: Unclear objectives cause a significant proportion of AI project failures. Projects without defined success metrics continue consuming budget without delivering measurable value. Failure thresholds prevent sunk cost fallacy where teams continue investing in projects that will never achieve returns. Gartner research predicts that 30% of generative AI projects will be abandoned after proof of concept by end of 2025, often due to unclear success criteria.

How to do it

  • Name the single primary KPI the AI project will improve (revenue, conversion rate, support ticket volume, processing time)
  • State the baseline metric, target metric, and timeline in one sentence
  • Define what “production ready” means in measurable terms (accuracy threshold, latency requirement, uptime target)
  • Set explicit go/no-go decision points at 25%, 50%, and 75% of timeline
  • Document what happens if the model performs below threshold (human review, fallback process, project cancellation)
  • Align all stakeholders on success definition before budget approval

Red flags to watch for

  • Different executives describe the project goal using different metrics
  • The business case assumes perfect model accuracy from launch
  • No fallback process exists for when the AI system fails or produces incorrect results
  • Project scope has changed more than twice during planning phase

Decision threshold: If stakeholders cannot agree on a single primary success metric in a 2-hour workshop, the project lacks sufficient clarity to justify budget commitment.


Step 6: Validate Budget Against Risk Profile

What it is: Cross-referencing the proposed budget against the risks identified in Steps 1 through 5 to determine if funding is adequate, excessive, or insufficient for the true scope.

Why it matters for budget decisions: Budgets built on assumptions rather than assessed risks consistently underestimate by 40 to 60%. This final validation step prevents approval of projects doomed to fail or cancellation due to overruns.

How to do it

  • Add data remediation effort from Step 1 to technical development budget
  • Include compliance and documentation effort from Step 2 in timeline
  • Factor hiring or external partnership engagement from Step 3 into team budget
  • Verify infrastructure budget from Step 4 covers 12 months of operations, not just development
  • Reserve 20% contingency for unknown risks in novel AI implementations
  • Compare total budget against expected ROI timeline (most AI projects require 12 to 18 months to show measurable returns)

Red flags to watch for

  • Budget assumes data is production-ready without remediation effort
  • Compliance obligations are treated as zero-effort activities
  • Team capacity gaps are ignored or assumed to resolve through “learning on the job”
  • No budget line exists for ongoing model retraining and maintenance

Decision threshold: If total validated budget exceeds initial estimate by more than 40%, return to project scoping before seeking approval. If ROI payback extends beyond 24 months, evaluate whether competitive advantage justifies the extended timeline.


When This Framework Changes

Rapid prototyping for single use cases: Teams building proof-of-concept prototypes with 6 to 8 week timelines and internal data only can use lightweight risk assessment. Full structured frameworks add overhead that delays learning. Switch to structured assessment before committing production budget.

Regulated industries with existing compliance programmes: Healthcare, financial services, and insurance SMBs with established ISO 27001 or ISO 22301 certification can accelerate Step 2. Existing compliance infrastructure handles 60 to 70% of AI-specific requirements.

Partnerships with certified AI engineering providers: SMBs engaging partners who maintain ISO certifications and demonstrate NIST AI Risk Management Framework alignment can delegate portions of Steps 3 and 4. The partner brings capability and infrastructure, reducing internal assessment burden. Partners like HST Solutions, which hold ISO 27001 and ISO 22301 certification, can embed senior AI engineers who bring both ML expertise and compliance readiness.

High-certainty AI applications: Narrow AI applications with proven track records in similar industries (fraud detection, demand forecasting, document classification) face lower technical risk. Increase focus on Step 1 data quality and Step 5 success metrics while streamlining capability assessment.


Real-World Decision Scenarios

Scenario: Financial Services SaaS Platform

Profile:

  • Company size: 120 employees
  • Revenue: €15M annually
  • Target market: European banks and insurers
  • Current state: Clean transaction data, no ML in production, SOC 2 certified
  • Growth stage: Series B, expanding to 5 new EU markets

Recommendation: Prioritise Step 2 regulatory mapping and Step 3 capability assessment before budget approval

Rationale: Financial services face high-risk AI classification under the AI Act when making credit or underwriting decisions. Existing SOC 2 certification provides compliance foundation but not AI-specific governance. Clean data removes Step 1 as primary blocker. The capability gap (zero production ML) and multi-market regulatory complexity require 3 to 4 months assessment before accurate budget estimates emerge.

Expected outcome: 4 months for risk assessment and partner selection, 6 months to production deployment, regulatory compliance documentation complete before launch

Scenario: Healthcare Technology Startup

Profile:

  • Company size: 35 employees
  • Revenue: €2M annually
  • Target market: UK and Irish GP practices
  • Current state: Patient data across 3 systems, no data governance, 1 data scientist
  • Growth stage: Seed funded, pre-revenue product

Recommendation: Complete Step 1 data quality remediation as separate project before AI budget commitment

Rationale: Healthcare data triggers both GDPR and sector-specific regulations. Three disconnected patient data systems guarantee data quality will block AI development. A single data scientist cannot handle data engineering, ML development, and healthcare compliance alone. Attempting AI development before data remediation wastes 60 to 80% of budget on rework.

Expected outcome: 9 months for data platform consolidation, 3 months for compliance assessment, 6 months for AI development once foundation exists

Scenario: B2B SaaS Analytics Platform

Profile:

  • Company size: 75 employees
  • Revenue: €8M annually
  • Target market: European enterprises across multiple sectors
  • Current state: 5 years of product usage data, 3 engineers with ML coursework, AWS infrastructure
  • Growth stage: Profitable, organic growth

Recommendation: Use structured framework but compress Steps 1 and 4, focus on Step 3 and Step 5

Rationale: Product usage data quality is likely high (generated by own system). AWS infrastructure exists. The gaps are production ML experience (coursework differs from deployment) and undefined success metrics. Spending 4 to 6 weeks on structured assessment reveals whether external partnership accelerates delivery or whether upskilling 1 to 2 engineers proves more effective long term.

Expected outcome: 6 weeks assessment, decision to hire senior ML engineer plus consulting engagement, production features in 5 months


FAQ

Q: How long should structured risk assessment take before committing AI budget?
Comprehensive structured assessment requires 4 to 6 weeks for SMBs with 50 to 200 employees. This includes data audits, regulatory mapping, capability assessment, and budget validation. Regulated industries add 2 to 3 weeks for compliance review. Rapid prototypes under 3 months can use lightweight assessment in 1 to 2 weeks.

Q: What percentage of AI budget should risk assessment consume?
Structured risk assessment typically consumes 5 to 8% of total AI project budget. This investment prevents the 40 to 60% budget overruns common in projects without upfront assessment. For projects exceeding 6 months delivery, assessment effort proves negligible compared to failure recovery.

Q: Should we assess risks internally or hire external consultants?
Assess internally when you have prior AI project experience and established compliance programmes. Engage external consultants when entering AI for the first time, operating in regulated industries, or when internal teams lack capacity to complete thorough assessment alongside existing responsibilities.

Q: How do European AI regulations change the risk assessment process?
The AI Act adds mandatory risk management and documentation for high-risk AI systems. European SMBs must assess regulatory classification before technical feasibility. GDPR requires Data Protection Impact Assessments for AI processing personal data. These obligations add 3 to 6 weeks to regulatory assessment compared to US-only operations.

Q: What happens if we skip risk assessment and start development immediately?
Teams skipping structured assessment face 80% project failure rates. Common outcomes include discovering data quality issues 4 months into development, uncovering regulatory requirements during vendor reviews, and hitting capability gaps when prototypes reach production. These discoveries force scope reduction, budget increases of 40 to 60%, or complete project abandonment.

Q: How often should we reassess risks during AI development?
Reassess risks at 25%, 50%, and 75% project completion milestones. Monthly reviews track whether identified risks are materialising or new risks are emerging. Infrastructure requirements should be reviewed monthly during initial production deployment. Regulatory environments change; the AI Act continues evolving, requiring quarterly compliance reviews for high-risk systems.

Talk to an Architect

Book a call →

Talk to an Architect