Upskilling Reality

    What Even Is an 'AI Engineer'? — 425 JDs and One Reader's Comment Later

    May 8, 2026·10 min read

    TL;DR

    49% of Software Engineer JDs already ask for GenAI. After scanning 425 fresh AI Engineer JDs, the role splits cleanly: 64% is backend engineers calling LLM APIs, 36% is agentic systems work. The label is doing more work than the boundary deserves — and the market knows it (less than 5% of JDs disclose salary).

    What "AI Engineer" actually means — banner

    A reader left a comment on my AI-vs-ML Engineer breakdown last week that I couldn't stop thinking about:

    "For over 50% of the market, 'AI Engineer' just means a backend engineer with an LLM API key."

    Reader comment pull-quote

    It was a throwaway line. It was also doing more work than the entire 2,000-word post above it.

    The reason it bothered me: I'd just published a comparison framing AI Engineer as a distinct role with a distinct salary band and a distinct skill stack. And here was a working engineer saying — quietly, in a comment — no, it's mostly just SWE who learned the OpenAI SDK.

    So I went back to the source.

    This week I pulled 1,101 fresh LinkedIn JDs in the US — 429 Software Engineer, 425 AI Engineer, 247 Machine Learning Engineer postings, all from the last 30 days. I wanted to know what "AI Engineer" actually means in the market that's currently hiring it.

    The commenter was more right than wrong.


    The 49% problem

    Start with what's happening on the SWE side.

    49% of Software Engineer JDs in the dataset mention GenAI in some form. 41% explicitly require AI/ML competence. The skills employers are starting to call out by name in SWE postings — Claude Code, Codex, OpenCode, "LLM-powered features," "RAG-based systems" — these are not aspirational signals from forward-looking companies. These are now baseline expectations for half of all SWE roles being posted right now.

    In other words: half the SWE market is already hiring for what an AI Engineer does in their day-to-day. Which raises the obvious question.

    If half of all SWE postings ask for GenAI, where exactly does the SWE role end and the AI Engineer role begin?

    💡Half the SWE market is already hiring for what an AI Engineer does day-to-day. The label "AI Engineer" is doing more work than the boundary deserves. The skill stack is the durable signal — the title is catching up to it.

    What the AI Engineer JDs actually ask for

    I read through 425 AI Engineer JDs and pulled out what employers explicitly require, what they assume as baseline, and what they ask candidates to add.

    The pattern:

    • 73% require Python
    • 22% require AWS
    • The most common ICP description in the JD bodies: "software developers transitioning into machine learning and AI model deployment, often with experience in programming and cloud infrastructure"
    • The top three "skills to add" called out across the dataset: prompt engineering, LLM integration, GenAI techniques

    If you read that list and squint, what you're looking at is a backend engineer's resume with three new bullet points on top.

    That's not a takedown — it's the actual hiring profile. And it lines up with what 64% of the JDs in the dataset are asking for: someone who can call LLM APIs cleanly, build RAG endpoints, ship LLM-powered features, integrate generative AI into existing software products.

    The other 36% is a different story. We'll get to it.


    The 64/36 split

    Pull out the agentic AI signal and the dataset cleanly bisects.

    The 64% block — what the title usually means: Backend engineer + LLM API + RAG. Python and AWS as foundation. Prompt engineering and LLM integration as the additive layer. The job is mostly: take a working software product and add an LLM-powered feature without breaking it.

    The 36% block — where the title earns weight: 155 of the 425 JDs explicitly require agentic AI work — autonomous agents, multi-agent orchestration, tool-using agents, agent-based architectures. This is a different role. It's systems work. It involves designing how agents communicate, how they fail safely, how state flows across tool calls, how multi-step reasoning gets evaluated. There's no SDK that abstracts it away yet.

    Both populations sit under the same job title. Both compete for the same career advice articles, the same bootcamps, the same LinkedIn feed.

    They are not the same job.

    What "AI Engineer" actually means — 64/36 split


    Three signals the role hasn't solidified

    If you're not sure which version of "AI Engineer" a posting is actually asking for, the JD itself usually won't tell you. Three signals will.

    Signal 1 — Less than 5% of AI Engineer JDs disclose salary. For comparison, 72% of SWE JDs in the same dataset disclose salary. Markets disclose comp when the market has agreed on what the role is worth. When the disclosure rate is this low, what you're looking at is a category that hiring managers are still defining on the fly.

    Signal 2 — The largest single hirer in the dataset is Prolific. $80/hour. Contract. 20 of the 425 JDs are from Prolific alone, all paying up to $80/hour, all contract. The next-largest hirer, Catalyst Labs, is also contract-flavored. Compare that to MLE, where General Motors leads with $170K–$300K full-time bands. When a category leans this heavily on contract and gig structures, the work is mostly project-shaped: ship a feature, evaluate a prompt, integrate a model. Real but bounded.

    Signal 3 — Only 36% of AI Engineer JDs require agentic AI work. That's the sliver where the title earns its weight. The other 64% is API integration with extra steps — important work, valuable work, but not categorically different from "backend engineer who learned the OpenAI SDK."

    ℹ️Less than 5% of AI Engineer JDs disclose salary, vs 72% for SWE. Markets disclose comp when they've agreed on what a role is worth. The number tells you the category hasn't settled.

    What MLE looks like by contrast

    To make the 64/36 split visible, contrast it with a role that has solidified — Machine Learning Engineer.

    In the same scan, 247 MLE JDs:

    Signal AI Engineer MLE
    Python required 73% 82%
    AWS required 22% 34%
    Deep-learning frameworks (PyTorch, TensorFlow) minimal 244 of 247 JDs
    MLOps required rare 34%
    Salary disclosure rate <5% 16%

    MLE hasn't been absorbed into the GenAI wave. The math, the training loops, the lifecycle management — these still gate on a different skill stack than "engineer who can call an LLM." It's the part of ML that didn't get commoditized when foundation models showed up.

    That's why the salary band stretches to $300K at the high end. Markets pay for skill stacks that haven't been easily reproduced. They underprice categories that any senior backend engineer can plausibly add to their resume in three months.


    Where to invest your upskilling time

    If you are a backend engineer already shipping LLM features at work — building a RAG endpoint, fine-tuning a prompt for a customer-support flow, integrating an LLM API into your product — you are already what 64% of the AI Engineer market is hiring for. You don't need a title change to confirm it. You need a portfolio that documents what you've shipped.

    If you want to differentiate — to land in the 36% rather than the 64% — three real frontiers, in order of how legibly they signal "this person is past the SDK layer":

    1. Agentic systems and multi-agent orchestration. Tool-using agents. Multi-agent workflows where state has to flow across calls. Failure modes — when does a tool call need a retry, when should an agent escalate, when does the system halt. This is where the JD market is starting to draw a real line, and the skill stack here doesn't reproduce in a weekend project. It takes shipping something agentic into production and watching it break.

    2. Evals and inference-time engineering. Most LLM features in production today have no rigorous eval pipeline. Hiring managers know this and are increasingly explicit in JDs about wanting candidates who can build one. Quantitative eval frameworks, regression suites for prompts, benchmark harnesses. Closer to test-engineering than to ML research, and a real differentiator.

    3. Fine-tuning and MLOps depth. Not because every AI Engineer needs to fine-tune a model, but because the engineers who can fine-tune are also the engineers who understand inference cost, latency, deployment pipelines, and observability. This is the skill stack that bridges into MLE territory if you ever want to make that jump.

    The common thread: they involve owning a system end-to-end, not just calling an API. That's the part the 64% job mostly doesn't ask for. That's why the 36% job is what most working professionals should actually be optimizing toward.


    Common mistakes engineers make on this transition

    ⚠️"AI Engineer" on your LinkedIn does not currently mean what it meant 18 months ago. The skill stack is the durable signal. The title catches up later. Pay for evidence of shipping, not for credentials.

    Chasing the title instead of the skill stack. "AI Engineer" means a different thing depending on which JD you're looking at. The skill stack is the durable signal. The title catches up later.

    Buying generalist AI courses to reproduce skills you already have. If you're a backend engineer with three years of API integration experience, a $4,000 LLM-Engineering course teaching you how to call an API is selling you something you can do already. Audit the syllabus before paying — if it stops at RAG basics, it's not where the gap is.

    Optimizing for "AI Engineer" pay numbers from inflated comp surveys. The actual disclosed salaries in the JD market are not the $200K+ figures you'll see in viral hiring infographics. The disclosure rate is 5%. The largest hirer is paying $80/hour contract. The pay band at the top of the market exists, but it's narrow, it's at the agentic-systems end, and it's gated on demonstrable systems work — not titles.

    Treating MLE and AI Engineer as a slope when they're a fork. A SWE doesn't gradually become an MLE by writing more LLM features. The math, training loops, and MLOps are gating skills, not optional ones. If you want MLE comp, you need to commit to that path. If you want AI Engineer access, you've probably already qualified.


    Where this leaves us

    The label "AI Engineer" is doing too much work right now. It's covering a fat 64% of API-integration work that any competent backend engineer can ramp into in 3–6 months, and a smaller 36% of agentic-systems work that requires deeper investment.

    The market hasn't agreed on which one the title points to. Until it does, the working professional's move is to ignore the label and pick the underlying skill stack.

    If you're already shipping LLM features, you're already inside the 64%. Document it, ship one more thing publicly, move on.

    If you want to be in the 36%, build something agentic and watch it break in production.

    The label is doing too much work. The work underneath is what gets paid.


    Source: LinkedIn JD scan — May 2026 (1,101 postings: 429 Software Engineer, 425 AI Engineer, 247 Machine Learning Engineer, US market, last 30 days). Cluster categorisation is LLM-assisted with mechanical re-counting on headline figures — treat percent shares as directional, not precise. · Dexity.com

    Dexity Sprint

    Ship Production Code with AI

    Most senior engineers have tried Cursor or Claude Code and ended up with larger PRs, more review cycles, and hidden technical debt.

    View sprint
    Abhinav Rawat

    Abhinav Rawat

    Co-Founder, Dexity

    Connect on LinkedIn
    Questions or suggestions?hello@dexity.com