Skip to main content
← Back to the Journal
Opinion · Project Management·May 2026·13 min read

Why AI projects fail in the Middle East.

In early 2025 we sat with a delivery lead in the Gulf who scrolled a spreadsheet labeled “AI initiatives.” Dozens of rows, most stuck between “pilot” and “re-scoping.” The shock was not the ambition — it was how few rows had a measurable production button a frontline employee presses every day.

This essay is not morality theater. We build and sell AI. But from our practice, the Middle East paid for slide velocity faster than it bought measurement charters, so failures look technical when they are usually organizational [1][2]. Five patterns — then a four-week gate and a decision grid you can reuse before the next renewal.

What we call failure by month twelve.

Our working definition: nobody uses the system, leadership cannot answer what changed in minutes or dirhams saved, or KPIs exist only in marketing copy. When we ask why AI projects fail in the Middle East, the answer is rarely “bad math on transformers.” It is missing owners, missing baselines, missing stop rules [3][5].

Pattern 1: PoC theater.

Vendors ship immaculate screenshots; your PDFs arrive with crooked scans and silent table breaks. The gap is not Arabic support — it is operable OCR, metadata, and policy on what may enter a model at all [6].

Our antidote: force a two-week pipeline on a bounded dirty sample before GPU shopping. If you cannot pass that gate, stop spending — read AI PoC theater and vendor demos alongside this piece.

A demo trained on cherry-picked cleanliness is not proof. Proof is the first thousand real requests on your own messy corpora.

Pattern 2: data that is not really there.

Enterprises buy “Arabic understanding” without owning field dictionaries, retention policies, or DPIA coverage. The sobering reality from our audits: a thin slice of files is model-ready; the rest needs human classification and legal sign-off before it ever becomes training or RAG corpus [3].

Pattern 3: contracts without internal owners.

A CTO signature without a product owner who can kill a release is how you get shadow usage, unaudited prompts, and procurement renewals driven by fear of looking late. Minimum viable governance: named application owner, weekly scoreboard, security sign-off on regulated paths [2][4]. For purchase discipline, use ten questions before signing an AI vendor.

Pattern 4: quarterly deck reboots.

Every quarter someone announces a new frontier nameplate. Without a frozen model line and three evaluation cycles, integrations never finish and your team burns on migration instead of product polish [5].

Pattern 5: goals with no numbers.

“Become more efficient” is not a KPI. Boards rightfully ask month twelve what moved. If you cannot quantify baseline handle time, defect rate, tickets closed, or SAR at risk, your GenAI program is a vibes line item [3].

Quick decision grid.

  • Data — red flag if two weeks on a dirty sample never started; green if OCR plus RAG ingestion works on a production-shaped slice.
  • Measurement — red if no numeric acceptance test pre-contract; green if one KPI updates weekly for product and compliance leadership.
  • Ownership — red if only the vendor has runbooks; green if your team can pause traffic and explain why.
  • Vendor cadence — red if slide cadence outpaces release notes; green if model changes require a written risk exception [5].

Four-week gate before the big signature.

  • Week 1: data inventory plus legal constraints checklist.
  • Week 2: pipeline on real records with human-labeled success examples.
  • Week 3: baseline KPI and manual control runbook.
  • Week 4: expand / stop / rewrite decision signed by product plus compliance.

Closing.

MENA does not need more hype. It needs charters: who owns data, who owns measurement, who can stop waste. We would rather kill a program in four disciplined weeks than fund twelve months of ambiguity.

If you are allocating budget this quarter, pair it with year-one LLM TCO — do not finance silicon because a deck looked brave. If your team cannot answer mission and metric on one page this week, you already know where the work starts.

Frequently asked questions.

  • Why do AI projects fail in the Middle East? Buyers often fund narrative before data readiness and KPIs, turning programs into repeated demos [2][3].
  • Is the model the root cause? Rarely alone; integration, compliance, and staffing gaps dominate post-hoc reviews [5][6].
  • How much time before production? Four weeks can validate a narrow high-value workflow if data exists; absent data, timelines stretch months [3].
  • Should we outsource everything? External help works with explicit metrics and exit rights — mirror Omani AI contract clauses.
  • We have ten stalled pilots — what now? Pick three high-value flows, cancel the rest for this quarter, and fund one with governance attached.

Sources.

[1] McKinsey — The State of AI (enterprise adoption series).

[2] Gartner — AI adoption and procurement risk materials.

[3] NIST — AI Risk Management Framework (AI RMF 1.0).

[4] OECD — OECD AI Principles.

[5] Nuqta — procurement retrospectives across Gulf deployments, May 2026.

[6] Nuqta — Arabic RAG readiness checklist, April 2026.

Related posts

Share this article

← Back to the JournalNuqta · Journal