hackcapital/Unsplash

Corporate leaders are racing to hire artificial intelligence talent, convinced that a few high-profile specialists can transform their businesses overnight. Instead, many are discovering that without strong data engineering foundations, those expensive AI teams are stuck waiting for clean, reliable information that never quite arrives. The result is a growing pattern of stalled projects, frustrated workers, and quiet course corrections as firms realize they have hired for the glossy front end of AI while neglecting the plumbing that makes it work.

I see the same story repeating across sectors: companies swap out data engineers for AI specialists, then wonder why their models fail in production or never leave the lab. The backfire is not that AI experts are unnecessary, but that they are being hired into environments that lack the data infrastructure, governance, and experienced workers needed to turn prototypes into durable products.

AI hiring booms while data engineering lags

Across corporate functions, job postings that mention artificial intelligence and machine learning are surging, from customer-facing teams to back-office operations. Roles in Sales, legal, engineering, marketing, and technology are all seeing greater availability for AI and ML positions, as executives try to embed automation and predictive tools into every corner of the business. The hiring pattern signals a belief that AI specialists can be dropped into existing teams to unlock value quickly, even when the underlying data systems are fragile or fragmented.

That enthusiasm is especially visible in the United States, where companies are hiring more AI specialists than data engineers despite the fact that AI cannot exist without robust data pipelines. In less tech-mature regions, the imbalance is even sharper, with organizations treating AI and data engineering as interchangeable rather than as two very distinct positions that require different skills and responsibilities. When firms prioritize algorithm talent over the people who manage ingestion, quality, governance, and infrastructure, they effectively build skyscrapers on sand, a choice that almost guarantees rework and disappointment once projects hit real-world complexity.

The skills mismatch inside AI teams

Inside many AI teams, the imbalance shows up as a quiet but costly skills gap. Most AI engineers are strong in model development and software design, comfortable tuning neural networks or integrating APIs, yet they often lack production-grade data expertise. I have seen teams where highly paid specialists spend weeks writing brittle one-off scripts to pull data from legacy systems, because no one with deep experience in scalable pipelines, orchestration, or observability was hired to support them. The result is a patchwork of ad hoc solutions that crumble under load or fail silently when upstream schemas change.

That gap is not a criticism of AI engineers, it is a reflection of how job descriptions are written and teams are staffed. When leaders assume that a single AI expert can also design storage layers, manage governance, and architect streaming systems, they ignore the warning that AI teams today lack production-grade data expertise and that this work cannot be solved by frameworks or architecture diagrams alone. Data engineering is its own discipline, with hard-won knowledge about batch versus streaming trade-offs, cost control in cloud warehouses, and how to keep personally identifiable information safe while still enabling analytics. Treating it as an optional add-on to AI work is a strategic mistake.

Why the infrastructure gap makes projects stall

The most visible symptom of this hiring imbalance is the growing number of AI initiatives that never make it past pilot stage. Models are trained on carefully curated datasets, perform well in controlled tests, then fail when exposed to messy, real-time production data that was never properly modeled or cleaned. In many organizations, there is no clear owner for data lineage, no standardized way to track how fields are transformed, and no reliable process for handling missing or corrupted records. Without data engineers to build and maintain that infrastructure, AI specialists are left trying to debug problems that originate far upstream of their code.

Industrial manufacturing offers a stark example of how this plays out. Industrial manufacturing has not commonly built organizational capability in data science, and while engineering roles are well established on factory floors, the expertise needed to create and maintain machine learning environments is not easily accomplished. When plants add sensors and Internet of Things devices, they generate torrents of data that must be ingested, normalized, and stored before any AI model can extract value. Without dedicated data engineers to design those pipelines, companies end up with siloed logs and dashboards instead of predictive maintenance systems or quality-control algorithms that actually work.

The human cost: layoffs, rehiring, and disillusionment

The misalignment between AI hype and operational reality is not just a technical problem, it is reshaping workforces in ways that are already proving unsustainable. Some employers have rushed to cut experienced staff in the belief that AI tools can replace them, only to discover that the promised productivity gains are slower and more fragile than expected. Inside Company cultureHR TechnologyTalent Management discussions, leaders are now confronting the risk that they have over-indexed on automation promises rather than proven technology, and that a portion of the workers they let go will need to be quietly rehired to keep operations running.

Analysts are also warning about an experienced worker crisis that will intensify as organizations try to scale AI without enough people who understand both the domain and the data. In Predictions for The Future of Work, Forrester Research has highlighted how this shortage is likely to accelerate across industries in 2026, as firms realize that they cannot simply plug in generic AI models and expect them to navigate complex regulatory, safety, or customer-service environments. When companies sideline veteran employees who know where data comes from, how it has been used, and what edge cases matter, they lose the institutional memory that makes AI outputs trustworthy and actionable.

Recruitment hype and the AI buzzword problem

The hiring imbalance is amplified by how recruitment itself has been swept up in AI buzz. AI in recruitment is a buzzword and all the industry leaders are talking about it, with vendors promising automated screening, instant matching, and predictive assessments that will surface the perfect candidate. I see job ads that sprinkle in terms like “generative AI” and “LLM” as if they were magic spells, even when the underlying role is closer to traditional analytics or software engineering. That marketing gloss encourages companies to chase AI-branded talent instead of carefully mapping the mix of data engineering, governance, and domain expertise they actually need.

At the same time, recruiters are under pressure to use AI tools to speed up hiring, which can create a feedback loop where algorithms trained on past job descriptions and resumes reinforce the very skills imbalance that is causing trouble. Guides on how to use AI in recruitment emphasize automation of resume parsing and candidate outreach, but they rarely address whether the models understand the difference between a data engineer and an AI specialist or can evaluate experience in building resilient pipelines. When hiring processes treat all technical roles as interchangeable “AI jobs,” organizations end up with glamorous titles and thin coverage of the unglamorous work that keeps systems running.

What data engineers actually do that AI specialists cannot replace

Part of the problem is that data engineering work is often invisible when it is done well. Data engineers design schemas, build ingestion jobs, manage ETL and ELT processes, and ensure that data is partitioned, indexed, and stored in ways that make queries fast and affordable. They handle governance, from access controls to retention policies, and they are the ones who get paged when a nightly batch fails or a streaming job falls behind. AI specialists may write clever models, but without this foundation, those models either starve for input or consume dirty data that leads to biased or unstable predictions.

There is also a misconception that AI itself will soon automate data engineering out of existence. In one widely shared discussion about whether AI will replace data engineers, a presenter on a platform like Replit explains how tools can automate coding tasks and generate boilerplate, yet still acknowledges that these systems have no idea how to interpret the messy, context-specific realities of enterprise data. The video, recorded in Aug, shows that while AI can help write SQL or Python, it does not understand the business rules, regulatory constraints, or operational edge cases that data engineers navigate every day. Rather than replacing them, AI tools are more likely to become part of their toolkit, speeding up routine tasks while leaving the hardest design and governance decisions to humans.

Sector-specific risks when AI outruns data maturity

The risks of hiring AI specialists into data-poor environments vary by industry, but the pattern is consistent. In manufacturing, as noted earlier, Industrial firms that have not commonly built organizational capability in data science struggle to connect their operational technology with modern analytics platforms. While engineering roles on the shop floor focus on physical equipment and process control, the work of integrating sensor streams into machine learning environments is not easily accomplished without dedicated data engineering. When companies skip that step, they end up with isolated pilots that never scale beyond a single line or plant.

In services sectors like finance, healthcare, and legal, the stakes are different but just as high. Sensitive information must be handled under strict compliance regimes, and any AI system that touches it needs carefully designed access controls, audit trails, and anonymization strategies. If organizations hire AI specialists to build chatbots or decision-support tools without also investing in data engineers who understand governance and infrastructure, they risk breaches, regulatory penalties, or flawed outputs that erode trust. The same pattern shows up in marketing and customer support, where teams rush to deploy generative AI on top of CRM data that has never been properly deduplicated or labeled, leading to embarrassing errors and customer frustration.

How companies can rebalance their AI and data strategy

Fixing this backfire starts with a more honest assessment of what AI projects actually require. Instead of treating AI specialists as silver bullets, leaders need to map the full lifecycle of their initiatives, from data collection and cleaning to deployment, monitoring, and iteration. That exercise almost always reveals a need for more data engineers, not fewer, along with clear roles for governance, security, and domain experts who can validate outputs. When I talk to teams that have successfully scaled AI, they describe a partnership model where data engineers, AI specialists, and business stakeholders work in lockstep rather than in silos.

Rebalancing also means changing how success is measured. Rather than counting how many AI roles have been filled or how many models have been trained, organizations should track metrics like data pipeline reliability, time to integrate a new source, or the percentage of AI features that are actually used in production systems. Guidance that urges leaders to stop hiring AI engineers and start hiring data engineers is not an argument against innovation, it is a reminder that sustainable AI depends on boring but essential infrastructure. When companies invest in that foundation, their AI specialists can finally focus on what they do best, and the promise of intelligent systems has a chance to move from slide decks to real-world impact.

More from MorningOverview