AI Engineers Wire APIs. ML Engineers Fine-Tune the Model.

For the last two years, the dominant story about AI Engineers has been that the job is mostly API plumbing — wire up an LLM, add RAG, ship the product. But there's a counter-narrative gaining traction: fine-tuning is making a comeback, and AI Engineers are the ones it's coming back for.
We pulled 6,802 job postings from the Skillenai index — 2,795 Data Scientist, 2,893 ML Engineer, and 1,114 AI Engineer roles posted between March and early May 2026 — to test that claim. The cross-section is striking. The time series isn't. And the most interesting finding is that "fine-tuning" turns out to mean two completely different things depending on which role posts the job.
The Headline: AI Engineers Mention Fine-Tuning Most
Across all 6,802 postings, fine-tuning prevalence by role looks like this:
| Role | Postings | % mentioning fine-tuning | 95% CI |
|---|---|---|---|
| Data Scientist | 2,795 | 4.9% | 4.1% – 5.8% |
| ML Engineer | 2,893 | 18.2% | 16.8% – 19.6% |
| AI Engineer | 1,114 | 26.2% | 23.7% – 28.9% |
An AI Engineer posting is more than five times more likely to mention fine-tuning than a Data Scientist posting, and roughly 1.4× more likely than an ML Engineer posting. All three pairwise gaps are highly significant (chi-square p < 1e-7 across the board, Cramér's V = 0.23). On the surface, the "fine-tuning is for AI Engineers" narrative looks correct.
Then You Look at What "Fine-Tuning" Actually Means
Inside each role, we counted how often a fine-tuning posting also mentioned ten adjacent skills, split into two clusters:
- Deep fine-tuning skills: LoRA, PEFT, RLHF, distillation, post-training, supervised fine-tuning (SFT)
- LLM-orchestration skills: RAG, prompt engineering, LangChain, vector databases
| Skill | DS (n=137) | MLE (n=526) | AIE (n=292) |
|---|---|---|---|
| Deep fine-tuning skills | |||
| LoRA | 9.5% | 8.9% | 6.8% |
| PEFT | 9.5% | 4.2% | 3.4% |
| RLHF | 10.2% | 8.4% | 7.2% |
| Distillation | 1.5% | 12.2% | 5.1% |
| Post-training | 2.2% | 6.8% | 4.5% |
| SFT | 6.6% | 8.0% | 2.7% |
| LLM-orchestration skills | |||
| RAG | 37.2% | 27.9% | 61.0% |
| Prompt engineering | 32.1% | 16.7% | 41.4% |
| LangChain | 21.2% | 8.9% | 33.9% |
| Vector database | 17.5% | 10.3% | 29.5% |
Read the AI Engineer column: every single deep fine-tuning skill is lower in AI Engineer postings than in ML Engineer postings. Distillation is more than twice as common in MLE fine-tuning postings (12.2%) as in AIE (5.1%). SFT is three times as common.
Now read the orchestration cluster: 61% of AI Engineer fine-tuning postings also mention RAG. 34% mention LangChain. 30% mention vector databases. These rates are 2–4× the ML Engineer rates.
The pattern is unmistakable. When an AI Engineer posting says "fine-tuning," it almost always says it as one item on a long list dominated by RAG, prompt engineering, LangChain, and vector databases. When an ML Engineer posting says "fine-tuning," it's far more likely to be paired with the actual technical machinery of fine-tuning — distillation pipelines, post-training, LoRA, RLHF.
Two Quotes From Real Postings
The split is even clearer in the prose. From an AI Engineer job description (NeuBird.ai):
"You will work with RLHF and fine-tuning of LLMs. This role requires a blend of analytical skills, creativity, and strong AI domain knowledge."
Note the framing: fine-tuning sits next to "creativity" and "AI domain knowledge." It's a competency to mention, not the core of the job.
From an ML Engineer job description at AMD:
"You have experience building and operating end-to-end ML pipelines for training, fine-tuning, reinforcement learning, and agentic systems."
Or even more starkly, a posting at Reflection AI literally titled Forward Deployed Engineer | LLM Post-training:
"Drive model fine-tuning and evaluations for enterprise customers. Run fine-tuning workflows on customer data."
For ML Engineers, fine-tuning is sometimes the entire job. For AI Engineers, it's a checkbox in a list dominated by orchestration.
Did Anything Move in Two Months?
Short answer: not in any way the data can confirm. We ran a Mann-Kendall trend test on the five weeks with at least 100 postings per role:
| Role | Range across 5 weeks | Mann-Kendall p |
|---|---|---|
| Data Scientist | 2.5% – 6.0% | 0.81 |
| ML Engineer | 13.9% – 21.1% | 0.46 |
| AI Engineer | 14.6% – 31.9% | 0.46 |
Eight weeks of data and roughly 150 AI Engineer postings per week is not enough to see a trend — week-to-week noise dwarfs any real movement. The ML Engineer series is the only one with a hint of directional shape, climbing from 13.9% in early April back to 21.1% by late April, but with five points it's a curiosity, not a finding. We'll revisit in Q3 when there's a longer time series.
So Is Fine-Tuning Being Absorbed by MLEs and Data Scientists?
By Data Scientists? Hardly. At 4.9% prevalence, fine-tuning is barely on the DS radar. When it does appear, it tends to be research-flavored — PEFT and LoRA show up at similar rates to MLEs — but the role is still anchored in SQL, statistics, and experimentation, not model surgery.
By ML Engineers? Partially yes — and this is the more interesting answer. The deep technical fine-tuning toolkit lives in MLE postings, not AIE postings:
- Distillation: 12.2% (MLE) vs 5.1% (AIE)
- Post-training: 6.8% (MLE) vs 4.5% (AIE)
- LoRA: 8.9% (MLE) vs 6.8% (AIE)
- RLHF: 8.4% (MLE) vs 7.2% (AIE)
If "fine-tuning is being absorbed by other roles" means the actual training-pipeline work — distillation, RLHF, post-training, LoRA — is concentrated in ML Engineer postings rather than AI Engineer postings, then yes. AI Engineers have the word "fine-tuning" more often. ML Engineers have the substance of it.
What This Means For Your Career
If you're an AI Engineer and you want fine-tuning to mean something more than "I called the OpenAI fine-tuning API once": treat the orchestration co-mentions as a warning. The role is shaped around RAG and LangChain, not LoRA and post-training. To deepen, you need to seek out postings that actually pair fine-tuning with distillation, post-training, or RLHF — and most of those postings carry the title "ML Engineer," not "AI Engineer."
If you're an ML Engineer: fine-tuning prevalence in your role is solidly in the high teens and the deep technical signal is concentrated here. Distillation, RLHF, and post-training are MLE differentiators in 2026. If you're hiring for a role that needs hands-on training-pipeline work, "ML Engineer" is the title that selects for it; "AI Engineer" is not.
If you're a Data Scientist: fine-tuning isn't expected of most DS roles, but the small slice of DS postings that do mention it tend to be the most methodologically deep — PEFT, supervised fine-tuning, preference alignment. If that's where your interest lies, target healthcare LLMs, recommendation systems research, and ML-research-flavored DS roles.
The cleanest one-line summary: AI Engineers fine-tune the API. ML Engineers fine-tune the model.
Methodology
Job postings pulled from the Skillenai prod-enriched-jobs index (8 weeks, 2026-03-10 → 2026-05-03). Roles identified by exact match on the parsed role field, with aliases collapsed for ML Engineer (ML Engineer + Machine Learning Engineer) and AI Engineer (AI Engineer + Artificial Intelligence Engineer). Fine-tuning prevalence measured by phrase match across fine-tuning, fine tuning, finetuning, fine-tune, finetune on the posting text. Statistical tests: chi-square contingency for cross-section, Mann-Kendall on weekly proportions for trend, Wilson score intervals for 95% CIs. Big Tech employers (Google, Apple, Microsoft, NVIDIA, Netflix) are under-represented because their proprietary ATS platforms are not scraped — this likely under-states the deep fine-tuning signal in MLE postings, supporting rather than threatening the headline finding. Full data and reproduction code: skillenai-notebooks/fine-tuning-by-role.