The Problem You're Actually Facing
You've posted a job listing for an AI engineer. It's been live for six weeks. You've received 200 applications. You've interviewed 15 candidates. None of them can actually ship production code.
This isn't an anomaly. This is the state of AI hiring in 2025.
The mismatch between what organisations need and what the talent market offers has created a crisis that goes deeper than simple supply-and-demand economics. According to the Stanford 2024 AI Index Report, the demand for AI skills has outpaced supply by a factor of 8:1 in some sectors. The World Economic Forum AI Talent Gap Article documents that this gap is widening, not closing. And yet, most hiring processes are still structured around assumptions that no longer hold.
The core issue isn't that AI talent doesn't exist. It's that the talent market has fractured into specialisations so narrow that most job descriptions are chasing unicorns—candidates who don't exist in sufficient numbers to meet demand. A person who can design large language model architectures, fine-tune models at scale, build retrieval-augmented generation systems, implement agentic workflows, secure production deployments against prompt injection, and navigate governance frameworks is extraordinarily rare. Yet many organisations expect to find exactly that person, at mid-market salaries, available immediately.
This article cuts through the noise. We'll map where the gap actually exists, why traditional hiring fails for AI roles, and—most importantly—how to build capability without waiting for the market to catch up.
Why Job Listings Don't Match Reality
Most AI engineer job descriptions are written by people who don't understand AI engineering. They're assembled by HR teams pulling from templates, reviewed by hiring managers who've never shipped a model to production, and posted by recruiters optimising for keyword volume rather than candidate fit.
The result is a job listing that reads like this:
"Seeking AI Engineer with 5+ years experience in machine learning, deep expertise in transformers, proven track record with LLMs, experience deploying to cloud infrastructure, knowledge of MLOps best practices, familiarity with vector databases, understanding of retrieval-augmented generation, experience with prompt engineering, background in computer science, strong software engineering fundamentals, experience with Python, Rust, and Go, familiarity with Kubernetes, demonstrated ability to work in fast-paced startups, experience in financial services or healthcare preferred."
That's not a job description. That's a fantasy. And it reveals a fundamental misunderstanding of how AI capability actually maps to roles.
The McKinsey State of AI Report found that organisations struggle most not with finding candidates with specific technical skills, but with identifying what skills actually matter for their specific problem. A candidate who's spent three years optimising transformer inference at a large language model company may be completely lost when asked to build a production agentic workflow for claims automation. A researcher with deep expertise in fine-tuning may have never deployed anything to production. A software engineer with strong DevOps chops might have no experience with model evaluation frameworks.
The problem compounds when you consider the actual distribution of AI talent:
The Research Tier: PhDs and postdocs who've published papers on attention mechanisms, scaling laws, or novel architectures. These people rarely want to build production systems. They want to push the frontier of what's possible. They're typically found at research labs, top-tier tech companies, or academia. Hiring them for production work is expensive and often results in frustration on both sides.
The Big Tech Tier: Engineers at OpenAI, Google DeepMind, Anthropic, Meta, and similar organisations who've shipped large-scale systems. These people are in high demand, well-compensated, and rarely leave. When they do, they typically move to well-funded startups or consulting firms, not mid-market enterprises.
The Practitioner Tier: Engineers with 2-5 years of hands-on experience building AI systems in production. This is where the real shortage exists. There simply aren't enough people in this tier. The field is too young. Most people who could be in this tier are still in the Research or Big Tech tiers, or they're working at well-funded startups.
The Transitioning Tier: Software engineers, data engineers, and data scientists who are moving into AI engineering. These people have strong fundamentals but lack production experience. They're trainable, but they require mentorship and support. Most organisations don't have the infrastructure to provide that.
The Bootcamp Tier: People who've completed AI bootcamps or online courses. Some of these candidates are genuinely capable. Many are not. The signal-to-noise ratio is terrible, and most hiring processes lack the sophistication to distinguish between them.
Your job listing is probably competing for people in the Practitioner Tier. There are maybe 5,000-10,000 of them globally who are actively available. You're competing against every other organisation trying to hire them, plus the venture capital ecosystem that's funding startups and offering equity upside. Your offer needs to be exceptional to win.
The Skills Gap vs. The Experience Gap
These are different problems, and conflating them will destroy your hiring process.
The skills gap is about specific technical knowledge: Can this person write Python? Do they understand how transformers work? Can they set up a vector database? These skills are teachable. A strong software engineer can learn vector databases in two weeks. A data scientist can understand transformer architecture in a month with good resources.
The experience gap is about production intuition: Can this person anticipate failure modes in a deployed system? Do they know how to design for latency constraints? Can they evaluate a model's performance in the context of business constraints? Can they navigate the messy reality of getting a system into production when requirements change, stakeholders disagree, and infrastructure breaks?
Most organisations optimise their hiring for the skills gap while ignoring the experience gap. They want someone who already knows all the tools they've chosen, speaks the same technology stack, and can contribute on day one. What they actually need is someone with production intuition who can learn their specific tools quickly.
This distinction matters because it changes how you should approach hiring. If you're hiring for the skills gap, you're looking for someone with specific technical credentials. If you're hiring for the experience gap, you're looking for someone who's shipped systems under pressure and learned from failures.
According to Harvard Business Review AI Talent War, organisations that succeed in AI hiring focus on identifying production experience, not credential collecting. They look for people who've shipped something, anything, and can articulate what went wrong and what they learned.
What Organisations Actually Need vs. What They're Hiring For
Let's be concrete. Most mid-market and enterprise organisations pursuing AI automation fall into one of three categories:
Category 1: AI Automation for Operations
You want to automate claims processing, customer onboarding, compliance checks, or similar workflows. You need someone who understands business process automation, can design agentic workflows, and can integrate with your existing systems. You need production intuition around latency (these systems often run in real-time), cost (inference costs matter at scale), and error handling (what happens when the model fails?). You need someone who understands that this isn't a research problem—it's an engineering problem. This person needs to understand AI agents vs chatbots: why the difference matters for ROI because the distinction determines whether your solution actually saves money or just looks impressive in demos.
What you're probably hiring for: A machine learning engineer with experience in NLP and model deployment.
What you actually need: A software engineer with strong systems thinking who's shipped at least one production automation system and understands the difference between a prototype and something that runs 24/7.
Category 2: AI Strategy and Governance
You need to figure out where AI creates value in your organisation, how to structure your AI capability, and how to govern AI systems responsibly. This requires someone who understands both technology and business, who can translate between engineering and executive teams, and who can navigate the messy reality of organisational change. This person needs to understand that AI-native companies don't have IT departments — they have AI departments and help you restructure accordingly. They need to understand AI-native vs AI-enabled: what's the difference and why it matters so you can make strategic choices about your organisation's direction.
What you're probably hiring for: A senior AI engineer or AI architect with strategic experience.
What you actually need: Someone who's worked in multiple organisations, seen different approaches, and can articulate trade-offs clearly. This might be a consultant, but it might also be a mid-level engineer who's been exposed to strategic thinking. You're hiring for judgment and communication, not just technical depth.
Category 3: AI-Native Product Development
You're building a product that's fundamentally AI-native—where the core value proposition depends on AI. This requires someone who understands how to build products where the model is the feature, where you need continuous evaluation and retraining, and where user experience is tightly coupled to model performance. This person needs to understand the distinction between AI consulting vs AI engineering: why the distinction matters because your organisation needs engineering, not advice.
What you're probably hiring for: A machine learning engineer with product experience.
What you actually need: Someone who's built an AI product end-to-end, understands the feedback loops between user experience and model performance, and can operate in high uncertainty. This is genuinely rare, and it probably requires either recruiting from a startup that's shipped an AI product or hiring someone junior with strong fundamentals and giving them ownership.
The pattern here is consistent: organisations are hiring based on credential and technical depth when they should be hiring based on production experience and judgment.
The Real Cost of Getting This Wrong
Hiring the wrong AI engineer is expensive in ways that go beyond salary.
Direct Costs:
- Salary and benefits for someone who can't deliver
- Recruiting costs to find a replacement
- Onboarding costs for the replacement
- Lost productivity during the ramp period
Indirect Costs:
- Damaged credibility with stakeholders when the hire doesn't work out
- Demoralised team if the hire creates friction
- Delayed projects while you're figuring out the hire isn't working
- Technical debt from decisions made by someone who didn't understand production constraints
Opportunity Costs:
- The projects you could have shipped if you'd hired correctly
- The market opportunity you miss while you're still figuring out your AI capability
- The competitive advantage you lose to organisations that move faster
According to Forbes AI Talent Shortage Article, organisations that hire poorly in AI typically spend 6-12 months before they realise the hire isn't working. By that point, they've already invested significantly in onboarding, project planning, and team integration.
The cost of hiring wrong is often higher than the cost of hiring an external partner to solve the problem while you build internal capability.
Why External Partners Can Bridge the Gap (And When They Can't)
Many organisations respond to the AI talent gap by hiring external partners—consultants, agencies, or specialised firms. This makes sense in some contexts and is a terrible idea in others.
External partners are useful when:
You need production expertise you don't have internally. If you've never shipped an AI system to production, bringing in someone who has done it 20 times is valuable. They know the failure modes, the gotchas, and the sequencing. They can compress your learning curve from 18 months to 6 months. At Brightlume, we work with organisations exactly at this stage—we help them ship their first production AI system in 90 days, which builds internal capability while delivering immediate value.
You need to move quickly and can't wait for hiring. If you have a time-sensitive opportunity or a strategic initiative that needs to launch, bringing in external expertise to accelerate is sensible. This buys you time to hire the right people internally.
You need specialised expertise for a specific problem. If you're building a compliance copilot and need someone with deep expertise in regulatory frameworks, financial services, and agentic workflows, that's a narrow skill set. Hiring a full-time employee for that might not make sense. Bringing in a specialist for the architecture phase makes more sense.
External partners are a bad idea when:
You're using them as a substitute for building internal capability. If you're hiring external partners indefinitely without building internal expertise, you're creating dependency and missing the opportunity to build competitive advantage. AI is becoming table stakes. You need internal capability.
You're expecting them to solve problems they can't actually solve. Some external partners are better at selling than shipping. They'll promise production-ready systems in 90 days but deliver prototypes that don't scale. Make sure you're working with partners who have a track record of shipping production systems, not just consulting engagements.
You're not learning from the engagement. If your external partner ships a system but your team doesn't understand how it works or how to maintain it, you've wasted money. The best external engagements are ones where knowledge transfer happens alongside delivery.
The right approach is usually hybrid: bring in external expertise to accelerate your first production deployment while simultaneously hiring and developing internal capability. This gives you the best of both worlds—you ship fast and you build long-term capability.
Building Internal Capability: The Alternative to Hiring Unicorns
If you can't hire the AI engineer you need, you need to build them. This requires a different approach than traditional hiring.
Start with Strong Fundamentals, Not Credentials
Look for software engineers with strong fundamentals: people who understand systems thinking, can write clean code, understand distributed systems, and have shipped production software. These people are more common than AI specialists, and they're trainable. A strong software engineer can learn transformers, vector databases, and agentic workflows in 3-6 months with good mentorship and resources.
This means your hiring bar should focus on:
- Ability to write clean, maintainable code
- Understanding of systems design and trade-offs
- Experience shipping production systems under constraints
- Intellectual curiosity and ability to learn quickly
- Communication skills (seriously, this matters more than you think)
Credentials like "5 years of ML experience" or "published papers on transformers" are nice but not necessary. A strong software engineer with 2 years of production experience will often outperform a PhD who's never shipped anything.
Invest in Mentorship and Knowledge Transfer
If you're hiring people without AI experience, you need to invest in their development. This means:
- Pairing junior people with senior people who can guide them
- Allocating time for learning (not just project work)
- Providing access to good resources (courses, papers, communities)
- Creating a culture where asking questions is valued
- Having people work on progressively harder problems as they develop capability
This is not a cost—it's an investment. A junior engineer who goes through a good mentorship programme will be more valuable to your organisation than someone you hired as a senior who doesn't fit your culture.
Bring in External Expertise Strategically
Use external partners to accelerate learning and validate approaches. This might mean:
- Bringing in a specialist for a 4-week architecture engagement to design your first agentic system
- Having your team pair with external engineers during the initial build phase
- Getting external review and feedback on your system design before you scale
- Learning how to evaluate models, set up monitoring, and handle failure modes from someone who's done it before
The key is that the external expertise should be transferring knowledge to your team, not just delivering a system. After the engagement, your team should understand how the system works and how to maintain and evolve it.
Structure Your AI Capability Around Problems, Not Roles
Instead of hiring "an AI engineer," think about the specific problems you need to solve:
-
Claims automation: You need someone who understands business process automation, can design workflows, and understands the constraints of your systems. This might be a software engineer with process automation experience, not an AI specialist.
-
Customer onboarding: You need someone who understands user experience, can design conversational interfaces, and understands how to handle edge cases. This might be a product engineer or a backend engineer with strong systems thinking.
-
Compliance and governance: You need someone who understands your regulatory environment, can design systems that are auditable and explainable, and understands how to implement guardrails. This might be a senior engineer with financial services or healthcare experience.
By framing roles around problems rather than titles, you expand the pool of candidates and you're more likely to hire people who actually understand your domain.
Measure Progress Against Production Milestones, Not Credentials
Stop evaluating AI capability based on degrees, certifications, or years of experience. Start evaluating based on:
- Can they design a system that meets your latency requirements?
- Can they evaluate a model's performance against your business metrics?
- Can they implement security controls that prevent prompt injection and data leaks? (See AI Agent Security: Preventing Prompt Injection and Data Leaks for specifics.)
- Can they design monitoring and alerting so you know when the system is degrading?
- Can they implement governance frameworks so your organisation can audit and explain AI decisions?
- Can they ship incrementally and iterate based on feedback?
These are the skills that matter. Everything else is secondary.
The Organisational Structure Problem
Hiring the right people is only half the battle. You also need to structure your organisation so that AI capability can actually develop.
Most organisations have a fundamental structural problem: AI is either siloed in a data science team that reports to IT, or it's scattered across different business units with no coherent strategy. Neither structure works.
The organisations that are winning at AI are restructuring around the reality that AI is becoming core to how they operate. This means:
AI Engineering is a First-Class Function
AI engineering should report to the same level as product, engineering, and operations. It shouldn't be a sub-team of IT or data science. It should be a core function with its own budget, its own hiring authority, and its own strategic accountability.
Cross-Functional Teams, Not Siloed Experts
Your AI team shouldn't be separate from your product and engineering teams. Instead, you should have cross-functional teams where AI capability is embedded alongside product and engineering expertise. This means your product managers understand what AI can and can't do, your engineers understand how to build AI systems, and your AI specialists understand your product and business constraints.
This is what we mean by AI-native companies don't have IT departments — they have AI departments. The entire organisation is structured around shipping AI, not around traditional IT functions.
Clear Governance and Accountability
You need clear governance frameworks that define who can make decisions about which models to use, how to handle data, how to monitor for bias and degradation, and how to comply with regulations. This needs to be documented and communicated, not just understood informally.
See AI Agent Security: Preventing Prompt Injection and Data Leaks and our broader work on enterprise AI governance for specifics on how to implement this.
Investment in Tooling and Infrastructure
You need good tooling: version control for models, experiment tracking, evaluation frameworks, monitoring systems, and security controls. Without this infrastructure, even great engineers will struggle. The cost of good tooling is a fraction of the cost of hiring people, and it multiplies the effectiveness of the people you do hire.
How to Evaluate Candidates When You Don't Know What You're Looking For
Most hiring managers in traditional organisations don't have strong intuition about AI capability. This makes it easy to hire the wrong person. Here's how to evaluate candidates more effectively:
Ask Them to Explain a Production Decision
Ask candidates to walk you through a decision they made in a production system: "You had to choose between model A and model B. How did you decide?" Listen for whether they talk about:
- Latency requirements and how each model performed
- Cost constraints and inference pricing
- Accuracy metrics and whether they mattered for the business
- Failure modes and what they'd do if the model degraded
- Monitoring and alerting
- Trade-offs and constraints
If they can articulate these trade-offs, they have production intuition. If they talk about model architecture or academic papers, they might be a researcher, not an engineer.
Ask Them About Failures
Ask candidates: "Tell me about a time when your AI system failed in production. What went wrong? What did you learn?" Listen for:
- Specific details about what failed and why
- What they did to fix it
- How they prevented it from happening again
- What they'd do differently next time
If they can articulate a failure and what they learned, they have production experience. If they say they've never had a production failure, they probably don't have production experience.
Ask Them to Evaluate a Specific Problem
Describe a specific problem you're trying to solve: "We want to automate claims processing. We get 10,000 claims per day. Each claim has 20-50 pages of documents. We need to classify claims, extract key information, and flag claims that need human review. Latency isn't critical—we can process overnight. Cost is critical—we need to stay under $0.50 per claim. How would you approach this?"
Listen for whether they:
- Ask clarifying questions about requirements, constraints, and success metrics
- Think about the problem systematically (data pipeline, model selection, evaluation, monitoring)
- Consider cost and latency trade-offs
- Think about failure modes and how to handle them
- Talk about iteration and how they'd learn from feedback
If they jump straight to "use Claude Opus 4" or "fine-tune a model," they're not thinking systemically. If they ask good questions and think through trade-offs, they're probably capable.
Ask Them About Your Specific Context
Ask candidates: "We're a financial services firm. We have strong governance requirements. We need to be able to explain every decision the AI system makes. We need to audit the system. What does that change about how you'd approach this problem?"
Listen for whether they understand:
- Regulatory constraints and why they matter
- The difference between explainability and interpretability
- How to design systems that are auditable
- How governance affects architecture decisions
If they can talk about these constraints intelligently, they've worked in regulated industries before. If they look confused, they probably haven't.
Involve Your Best Technical People in Interviews
Don't rely on HR or hiring managers to evaluate technical capability. Have your best engineers interview candidates. They'll be able to distinguish between someone who can actually ship and someone who's good at talking about AI.
The 90-Day Path to Production
Once you've hired the right people (or brought in external partners), you need a structured approach to getting to production quickly. This is where the rubber meets the road.
At Brightlume, we help organisations ship production AI in 90 days. This isn't magic—it's a structured approach that focuses on outcomes, not perfection. Here's the framework:
Month 1: Discovery and Architecture
- Define the specific problem you're solving and success metrics
- Map the current process and identify where AI adds value
- Evaluate model options (Claude Opus, GPT-4, Gemini, etc.) against your constraints
- Design the system architecture, including data pipelines, model integration, and monitoring
- Set up the technical infrastructure (version control, experiment tracking, evaluation frameworks)
- Build a minimal viable prototype to validate assumptions
Month 2: Build and Evaluate
- Implement the core system
- Build evaluation frameworks to measure performance against business metrics
- Iterate on the model and system design based on evaluation results
- Implement security controls and governance frameworks
- Set up monitoring and alerting
- Begin integration with your production systems
Month 3: Deploy and Handoff
- Deploy to production (often starting with a limited rollout)
- Monitor system performance and handle issues
- Train your team on how to maintain and evolve the system
- Document the system, the decisions made, and the lessons learned
- Plan for the next phase of capability building
This approach works because it focuses on shipping, not perfection. You're not trying to build the perfect system. You're trying to build a system that works well enough to create value, can be monitored and improved, and is maintainable by your team.
For more on this approach, see 7 Signs Your Business Is Ready for AI Automation and AI Automation Maturity Model: Where Is Your Organisation?.
Specific Strategies for Different Roles
The hiring challenge looks different depending on what role you're trying to fill. Here are specific strategies for the most common roles:
Hiring an AI Strategy Lead or Head of AI
This person needs to understand both technology and business, and they need to be able to navigate organisational change. Look for:
- Someone who's worked across multiple organisations and seen different approaches
- Someone who can articulate trade-offs clearly
- Someone who understands governance and compliance
- Someone who's built AI teams before
This might be someone from a large tech company who's tired of working on massive scale problems and wants to focus on impact. It might be a consultant who's worked with multiple organisations. It might be someone from a startup who's shipped a product. The key is production experience and the ability to think strategically.
Salary expectations: This is a senior role. Expect to pay $150k-250k+ depending on experience and location. You might need to offer equity if you want someone from a successful startup.
Hiring an AI Engineer (Production Focus)
This person needs to ship systems that work at scale. Look for:
- Someone with 2-5 years of production experience
- Someone who's shipped at least one system end-to-end
- Someone who understands systems design and trade-offs
- Someone who can write clean, maintainable code
You can find these people at startups, at large tech companies (though they're more expensive), or by developing them internally from strong software engineers.
Salary expectations: $120k-180k depending on experience and location. In competitive markets, you might need to go higher.
Hiring a Data Scientist Transitioning to AI Engineering
Data scientists have strong fundamentals in statistics and modelling, but they often lack production experience. If you hire someone in this category, you need to:
- Pair them with a senior engineer who can mentor them on production systems
- Give them ownership of a project from day one, but with support
- Invest in their development (courses, mentorship, time to learn)
- Be patient—it'll take 6-12 months for them to be fully productive
This is a good strategy if you can't hire a senior AI engineer. You're making a longer-term investment in capability.
Salary expectations: $100k-150k depending on experience.
Hiring a Software Engineer to Transition into AI
Strong software engineers are easier to hire than AI specialists, and they're trainable. If you hire someone in this category:
- Make sure they have strong fundamentals (systems thinking, clean code, production experience)
- Invest heavily in their AI education (courses, mentorship, working on AI problems)
- Give them progressively harder problems as they develop capability
- Have them work alongside someone with AI expertise
This is the most scalable approach if you're building long-term AI capability. You're investing in people with strong fundamentals and developing them into AI engineers.
Salary expectations: $100k-150k depending on experience.
What Not to Do
Here's what doesn't work:
Don't Hire Based on Credentials Alone
A PhD in machine learning doesn't mean someone can ship production systems. A bootcamp graduate might be more useful than a PhD if they've shipped something. Evaluate based on production experience, not credentials.
Don't Hire Someone Senior for a Junior Role
If you hire a senior AI researcher to work on operational automation, they'll be bored and frustrated. You'll be paying senior money for junior work. Hire for the role you actually have, not the title you want to put on the job posting.
Don't Expect Someone to Be Productive on Day One
AI engineering is complex. Even someone with relevant experience will need 4-8 weeks to understand your systems, your constraints, and your business context. Plan for a ramp-up period. If you need someone to be productive immediately, you probably need external help.
Don't Hire in Isolation
If you hire an AI engineer but don't have the infrastructure, governance, or organisational structure to support them, they'll struggle. You need to hire as part of a broader strategy, not as a one-off hire.
Don't Use Hiring as a Substitute for Strategy
If you don't know what problems you're trying to solve with AI, hiring an AI engineer won't help. You'll end up with an expensive person looking for problems. Get clear on your AI strategy first, then hire to execute on that strategy.
The Path Forward
The AI talent gap is real. It's not going away anytime soon. But it's not insurmountable if you approach it strategically.
Here's what we recommend:
First, Get Clear on What You Actually Need
Before you post a job listing, get clear on the specific problems you're trying to solve. What business process do you want to automate? What constraints do you have (latency, cost, governance)? What does success look like? Once you're clear on these questions, you can hire for the right role.
Second, Evaluate Your Options
Do you need to hire? Could you solve this problem faster with external expertise? Could you develop capability internally over time? What's the right mix of hiring, external partnership, and capability building for your organisation?
If you're not sure, talk to organisations that have done this before. See Brightlume's case studies for examples of how different organisations have approached this.
Third, Hire for Fundamentals and Production Intuition
Stop looking for unicorns. Look for people with strong fundamentals (systems thinking, clean code, production experience) and production intuition (they've shipped something and learned from failures). These people are more common, and they're trainable.
Fourth, Invest in Capability Building
Whether you hire internally or bring in external partners, invest in building capability. This means mentorship, good tooling, clear governance, and organisational structure. This is a long-term investment, but it's the only way to build sustainable AI capability.
Fifth, Focus on Shipping, Not Perfection
Your first AI system doesn't need to be perfect. It needs to work, it needs to create value, and it needs to be maintainable. Ship something, learn from it, improve it. This is how you build capability.
For more on how to execute on this strategy, see AI Consulting vs AI Engineering: Why the Distinction Matters. The distinction matters because it determines whether you're getting advice or delivery.
The organisations that are winning at AI aren't the ones with the most impressive credentials. They're the ones that are shipping, learning, and iterating. They're the ones that have structured their organisations around AI capability, not around traditional IT functions. And they're the ones that understand that hiring is just the first step—building capability is the long game.
The AI talent gap is a real problem. But it's solvable if you approach it strategically, hire for the right things, and invest in building capability. Start there.