How to Choose the Right AI Consulting Firm: 10 Criteria for Enterprises

Boris Friedrich
Boris Friedrich
15 min read
How to Choose the Right AI Consulting Firm: 10 Criteria for Enterprises

Choosing the right AI consulting firm is not about brand name but ten criteria: regulatory fluency (EU AI Act, GDPR, ISO 42001), industry depth, a real methodology, vendor-independence, MLOps maturity, measurable references, mixed consultant-plus-engineer teams, transparent pricing, data protection, and post-project support. Provider selection is the single biggest risk in an AI programme.

Why provider selection — not technology — is the #1 risk in AI projects

In four out of five failed AI projects, the root cause is not the technology but the choice of partner. The model, the cloud, the LLM provider — all of that is replaceable. The consulting partner is not. An AI consulting firm makes decisions in the first few weeks about data architecture, operating model, and regulatory posture that are expensive to undo later. Pick the wrong firm and you lose not just budget but months of time-to-market — in a 2026 regulatory environment where every month counts.

This guide is not a pitch, it is a scoring rubric. The ten criteria are deliberately framed so you can apply them to any provider — including ADVISORI. At the end you will find red flags, a 5-step RFP playbook, and a build-vs-buy decision matrix for the question every executive sponsor asks first.

The 10 criteria — at a glance

  1. Regulatory fluency: EU AI Act, DORA, GDPR, ISO 42001
  2. Industry depth in regulated sectors: banking, insurance, healthcare
  3. Methodology: a real roadmap framework, not ad-hoc consulting
  4. Technology neutrality: vendor-independent, not cloud or LLM lock-in
  5. MLOps and production capability, not a PoC factory
  6. References with measurable outcomes — case studies with numbers
  7. Team composition: consultants PLUS engineers, not consultants only
  8. Transparent pricing: fixed-price phases, not open time-and-materials
  9. Data protection and on-premise capability — critical for DACH / EU
  10. Post-project support via a defined retainer model

Criterion 1: Regulatory fluency (EU AI Act, DORA, GDPR, ISO 42001)

The EU AI Act (Regulation (EU) 2024/1689) regulates AI by risk class and touches every enterprise that uses AI in customer-facing journeys, HR decisions, or critical infrastructure. For financial services, DORA adds ICT risk-management obligations. GDPR sits across everything. ISO 42001, in force since 2024, is the new management-system standard specifically for AI.

What this means for vendor selection: ask concretely who on the team has read the AI Act (not "is familiar with it"), who has taken a high-risk system compliantly into production, and how governance artefacts (risk assessments, datasheets, monitoring) are embedded in the delivery plan. A vendor who says "we handle regulatory later" has disqualified themselves — "later" means rework and audit findings.

Probe question: "Show me a use case you classified as high-risk under the AI Act and then ran compliantly in production — with what governance setup?" Abstract answers mean they have never done it.

Criterion 2: Industry depth in regulated sectors

AI consulting for a bank is not AI consulting for an e-commerce shop. In regulated sectors — financial services, insurance, healthcare, energy, public administration — the real bottleneck is not the model but data availability, audit trails, and approval workflows. A firm without industry depth optimises for model accuracy and is surprised when the regulator, the information-security officer, or the data-protection officer blocks deployment.

Industry depth shows up not on the logo wall but in the vocabulary. Ask about MaRisk, VAIT, or BAIT experience if you are a bank in DACH. Ask about MDR if you are in medical devices. Ask about NAIC principles if you are a US insurer. Fluency with these terms correlates almost perfectly with whether they have actually shipped in that regulated environment.

Criterion 3: Methodology — a roadmap framework, not ad-hoc consulting

Serious AI consulting is methodical. It follows a framework that explicitly names phases, artefacts, and decision points — typically Discovery, Pilot, Scale, Operate. Ad-hoc consulting delivers workshops and strategy decks but no verifiable interim outputs. A real framework reveals itself when the vendor shows you the artefact set per phase up front: which deliverables do you get, at what moment, at what quality bar?

Pay special attention to the transitions. A firm that hands you a slide deck between pilot and production creates PoC graveyards. A firm that hands you a scale artefact with data pipelines, monitoring dashboards, and runbooks knows how production actually works.

Criterion 4: Technology neutrality

Some consultancies are de facto sales arms of a cloud or model provider. That is not necessarily bad — but it must be disclosed. When a firm claims technology independence, insist on proof: show me two reference projects on different hyperscalers and one on-premise. A genuinely vendor-neutral partner can do this without flinching.

Lock-in is rarely created by the vendor alone, it is created by the architecture choices they recommend. Ask concretely: "Which components of our setup would be migratable to a different provider or stack in 24 months — at what cost?" A credible consulting partner has this answer before the project starts.

Criterion 5: MLOps and production capability

The biggest gap in the AI consulting market sits between proof-of-concept and production. Many firms are excellent at building an impressive prototype in four weeks — and leave the client with a Jupyter notebook that nobody can operate. MLOps maturity is the difference between "AI experiments" and "AI in the business".

Concretely: does the firm have model versioning, data and feature stores, monitoring for drift and data quality, automated retraining pipelines, and incident-response runbooks for AI systems? If any of these answers is hesitant, you will be walking the road to production on your own.

Criterion 6: References with measurable outcomes

A real AI reference contains three numbers: the baseline, the outcome, and the elapsed time. "We automated claims handling for an insurer" is not a reference. "We auto-triaged 68 percent of incoming claims, reduced handling time from 6 to 2 days, and shipped in 11 months" is one.

If the vendor cannot share numbers because of an NDA, they should still name the range and — ideally — arrange a reference call with the client-side programme owner. A firm that cannot make any reference reachable either has none or their clients refuse to speak. Both are signals.

Criterion 7: Team composition — consultants plus engineers

Pure advisory houses deliver slides. Pure implementation houses deliver code without a strategic frame. An AI project needs both in one team. In the sales conversation, ask who will actually sit on the delivery team: what share of senior consultant time? who writes the production code? who runs the system after go-live? is that one person, two, or fifteen?

A mixed team covering three to five roles — strategy consultant, data engineer, ML engineer, governance specialist, solution architect — spans what a serious AI programme needs. Firms offering less outsource parts of the work — ask them to whom.

Criterion 8: Transparent pricing

The clearest split in the AI consulting market runs through the pricing model. Open-ended time-and-materials (T&M) means an open budget with a built-in incentive to extend the project. Fixed price per phase with clear acceptance criteria distributes risk fairly: the vendor carries delivery risk, you carry scope risk. Hybrid models — fixed price for Discovery and Pilot, capped T&M for Scale — are usually the best of both.

Watch for hidden cost components: licences for proprietary libraries, subscription fees for model APIs, data-labelling costs on top of the day rate. A transparent vendor shows you the total cost of ownership over 24 months, not just the day rate for the next eight weeks.

Criterion 9: Data protection and on-premise capability

In the DACH region — and across EU financial services, healthcare, and public administration — the question "where does the data sit?" is not a technical footnote, it is a go/no-go criterion. A consulting firm that only knows US-hyperscaler setups will fail your data-protection review.

Probe questions: can the firm propose a setup where personal or sensitive data never leaves an EU data centre? Is there experience with European cloud providers (OVHcloud, Open Telekom Cloud, IONOS, STACKIT) or fully on-premise deployments? How is LLM inference handled — dedicated EU instances, private endpoints, self-hosted open-source models? A vendor who can answer concretely saves you months of architectural debate.

Criterion 10: Post-project support and retainer models

AI systems degrade in production — models drift, data distributions shift, compliance requirements tighten. A good AI consulting partner therefore does not stop at go-live. Ask explicitly about the model for the phase after: is there a retainer covering monitoring, model refresh, and minor extensions? what does it cost per month? who is the named point of contact?

The alternative — "call us when something breaks" — is not acceptable for a production AI system. A defined operating model after go-live belongs in the proposal, not in follow-up negotiations once the project budget is spent.

Red flags: how to spot a weak AI consulting firm

A few patterns show up in vendor conversations again and again. Take every one seriously:

  • Pure sales presentation without technical depth: after 60 minutes you still do not know how they work, but the day rate is on the table.
  • "Agentic AI" as a central term without clarifying what will actually be built — buzzword architecture instead of problem analysis.
  • A single technology stack in the portfolio (only OpenAI, only AWS, only Azure) — the firm recommends what it can, not what fits.
  • No concrete references with numbers, just a logo wall of companies where "something was done once".
  • Unclear split between delivery team and account team: seniors in the pitch, juniors on the project.
  • Regulatory work pushed into "phase two" instead of framing the project from day one.
  • No statements on MLOps, monitoring, or runbooks — the vendor plans to demo day, not to production.
  • A surprisingly low price well below market — the firm subsidises the entry project to lock you in at Scale.

How to run an RFP for AI consulting — 5 steps

  1. Summarise problem and outcome in a one-page brief: which business case, which KPI, what time horizon, what regulatory guardrails. Without this page, every RFP turns mushy.
  2. Invite three to five firms — fewer than three gives you no meaningful comparison, more than five wastes everyone's time. Mix deliberately: one big name, one specialist, one challenger.
  3. Require a solutioning workshop with real (anonymised) data or process samples. Slide-only responses drop out. Real firms show you in 2 to 4 hours how they think.
  4. Score proposals against the 10 criteria, with explicit weighting. Regulatory, methodology, and MLOps deserve more weight than price — a cheap firm without production capability is the most expensive in the end.
  5. Start with a time-boxed Discovery/Pilot (6 to 10 weeks) — only then tender Scale. This lets you test a firm with contained risk before the main budget flows.

The 5 questions to ask in a first call

  1. "Show me a use case you took to production in the last twelve months — with numbers on outcome and duration."
  2. "How would you classify our use case under the EU AI Act risk tiers, and which governance artefacts will the project produce?"
  3. "Which people in the room today will actually sit on the delivery team — and at what share of their time?"
  4. "Which components of our setup would be migratable to a different vendor or stack in 24 months — at what cost?"
  5. "What does the operating model look like after go-live — retainer, incident response, model refresh — and what does it cost monthly?"

Internal team vs external consultant: a decision matrix

Build-vs-buy for AI is rarely binary. Six factors push the answer one way or the other:

  • Strategic relevance: is AI a competitive differentiator for you (→ keep capability in-house) or standardisation (→ buy externally)?
  • Regulatory density: the more regulated, the more internal ownership you need — consulting can prepare, not replace.
  • Time pressure: an internal build takes 12 to 24 months. For acute needs, external expertise is the only realistic route.
  • Volume: below five productive use cases, an internal team does not amortise — the fixed cost of an AI team starts at three to five FTE.
  • Data readiness: are your core data sets available and usable in-house? if not, you need external data-engineering capacity first.
  • Talent market: can your location realistically hire senior ML engineers and governance profiles? if not, a hybrid model typically wins.

The most stable model for mid-sized DACH and EU enterprises, in our experience, is hybrid: a small internal team (2 to 4 people) owning strategy, product, and governance — flanked by external consulting for discovery, engineering peaks, and regulatory depth. You keep decision authority in-house and buy targeted capacity and expertise when you need them.

Frequently asked questions about choosing an AI consulting firm

What is the single most important criterion when choosing an AI consulting firm?

In regulated sectors, regulatory fluency (EU AI Act, GDPR, ISO 42001, sector-specific supervision) is by far the most important factor. A technically strong firm without regulatory depth will ship a system your data-protection officer or supervisor ultimately refuses to approve. In less regulated sectors, MLOps maturity and methodology move to the top of the list.

How many AI consulting firms should I invite to an RFP?

Three to five is the sweet spot. Fewer than three gives you no meaningful comparison; more than five burns everyone's time without improving the decision. Mix deliberately: one big name (benchmarking), one specialist (depth), one challenger (fresh perspective).

How long does a typical AI consulting RFP take?

For a mid-sized AI project plan on 6 to 10 weeks: 2 weeks for briefing and distribution, 3 to 4 weeks for proposals and solutioning workshops, 1 to 2 weeks for scoring and decision, 1 to 2 weeks for contract. Faster is only possible if your vendor shortlist is already pre-qualified.

Should we build AI in-house or hire an external consultant?

It depends on six factors: strategic relevance, regulatory density, time pressure, use-case volume, data readiness, and talent market. For mid-sized enterprises, a hybrid model usually wins — a small internal team owning strategy and governance, plus external consulting for engineering peaks and regulatory depth.

How do I spot a weak AI consulting firm?

Typical red flags: pure sales pitch without technical depth, buzzword stacks with no concrete architecture ("agentic AI"), no references with numbers, blurred split between pitch and delivery team, regulatory work treated as an afterthought, missing MLOps capability, and suspiciously low entry prices used as a lock-in lever.

Is a specialist or generalist AI consulting firm better?

For regulated industries, a specialist with sector depth is almost always superior — BaFin, MaRisk, or MDR experience is not something you acquire in passing. For broad transformation programmes spanning many business units, a generalist with a strong partner network may fit better. The pivotal question: is your regulatory exposure high or low?

How does the EU AI Act affect provider selection?

Centrally. The regulation entered into force in August 2024 and its obligations phase in through 2026 and 2027, with full effect for high-risk systems. An AI consulting firm that cannot classify your use case under AI Act risk tiers and does not include governance artefacts (conformity declaration, risk assessment, monitoring setup) in the proposal is not sufficiently qualified for projects with supervisory exposure.

What should an AI consulting contract cover?

At minimum: clear deliverables per phase with acceptance criteria; IP and code ownership after project end; a GDPR-compliant data-processing agreement; named key people on the delivery team; an escalation path; an exit clause with handover package; a retainer or support option after go-live; a liability clause for regulatory findings; and a clause governing subcontractors. Boilerplate consulting agreements missing these points are insufficient.

Hat ihnen der Beitrag gefallen? Teilen Sie es mit:
Further reading

Continue exploring with related insights from our experts.

Your strategic success starts here

Our clients trust our expertise in digital transformation, compliance, and risk management

Ready for the next step?

Schedule a strategic consultation with our experts now

30 Minutes • Non-binding • Immediately available

For optimal preparation of your strategy session:

Your strategic goals and challenges
Desired business outcomes and ROI expectations
Current compliance and risk situation
Stakeholders and decision-makers in the project

Prefer direct contact?

Direct hotline for decision-makers

Strategic inquiries via email

Detailed Project Inquiry

For complex inquiries or if you want to provide specific information in advance