AI Roles & Career Paths: The New Engineering Landscape
AI Roles & Career Paths: The New Engineering Landscape
Section titled “AI Roles & Career Paths: The New Engineering Landscape”Last updated: March 2026
Confidence: Tier 2 — Based on job market data, industry publications, and emerging field research
Reading time: ~20 minutes
The AI wave didn’t just create new tools. It created new jobs that didn’t exist 3 years ago and is reshaping what existing roles mean. This guide maps the full landscape: what each role does, what skills it requires, how they relate to each other, and where each one is heading.
Table of Contents
Section titled “Table of Contents”- The Landscape in One View
- Prompt Engineer
- Context Engineer
- AI Engineer
- LLM Engineer
- AI Agent Engineer
- Founding AI Engineer
- AI Architect
- Platform Engineer (AI context)
- Harness Engineer
- AI Product Manager
- AI Safety & Eval Engineer
- ML Engineer
- Career Decision Matrix
- Salary Benchmarks (2025-2026)
- What’s Not a Role (Yet)
- Job Listings
1. The Landscape in One View
Section titled “1. The Landscape in One View”Two axes structure this landscape: proximity to the model (are you training it, prompting it, or building infrastructure around it?) and proximity to production (research vs. shipped product).
← Closer to the model Closer to infrastructure →
Research ML Engineer ←────────────────────────── AI Architect AI Safety Engineer Platform Engineer │ │ │ │Production LLM Engineer ──── AI Engineer ──── AI Agent Engineer Context Engineer Harness Engineer Prompt Engineer Founding AI Engineer AI Product ManagerMost new job demand sits in the bottom-right: building reliable AI systems that ship and stay reliable in production. The “pure research” quadrant remains competitive and specialized. The highest growth is in the applied, product-facing roles.
2. Prompt Engineer
Section titled “2. Prompt Engineer”Status: First wave (2022-2023), partially commoditized but still relevant in specialized contexts.
What they do
Section titled “What they do”Craft and optimize the instructions sent to AI models to get reliable, high-quality outputs. The scope ranges from one-shot prompts to complex multi-step prompt chains for production systems.
Responsibilities
Section titled “Responsibilities”- Design prompt templates for specific use cases (customer support, code generation, document analysis)
- Run systematic A/B tests to measure prompt performance
- Document prompt libraries and version them
- Optimize prompts for cost (fewer tokens, same quality)
- Work with domain experts to encode knowledge into prompts
Required skills
Section titled “Required skills”| Technical | Soft |
|---|---|
| Understanding of LLM behavior and failure modes | Communication with non-technical stakeholders |
| Basic Python (for automation and testing) | Systematic experimentation mindset |
| Familiarity with evaluation frameworks | Attention to edge cases |
| Versioning practices | Documentation discipline |
Where it’s heading
Section titled “Where it’s heading”The “prompt engineer” title as a standalone role is consolidating into broader AI Engineer or Context Engineer roles. Where it persists: companies with very specific, high-stakes prompt domains (legal, medical, financial compliance). Upskill toward context engineering or AI engineering if you’re in this role.
Entry paths
Section titled “Entry paths”Technical writer, QA engineer, domain expert (law, medicine, finance), content strategist.
3. Context Engineer
Section titled “3. Context Engineer”Status: Emerging — one of the fastest-growing specializations in 2025.
What they do
Section titled “What they do”Context engineering is the evolution of prompt engineering. Where prompt engineers craft instructions, context engineers design systems that give AI models the right information, at the right time, in the right format. Andrej Karpathy explicitly moved from “vibe coding” framing to “context engineering” as the more precise description of this work.
“Context Engineering is providing the right information and tools, in the right format, at the right time.” — Philipp Schmid, Google
Responsibilities
Section titled “Responsibilities”- Design RAG (Retrieval-Augmented Generation) systems and knowledge bases
- Manage context windows across multi-turn interactions and long-horizon tasks
- Define what agents remember, retrieve, or forget during task execution
- Structure information hierarchies (system prompts, conversation history, retrieved docs, tool definitions, safety constraints)
- Optimize context for accuracy and cost simultaneously
- Measure context quality through systematic evals
Required skills
Section titled “Required skills”| Technical | Soft |
|---|---|
| Python (context pipeline automation) | Systems thinking |
| Vector databases (Pinecone, Chroma, Weaviate) | Information architecture instinct |
| SQL and NoSQL (context retrieval) | Cross-functional collaboration |
| Cloud platforms (AWS/Azure/GCP) | Curiosity and continuous learning |
| RAG architectures, embedding models | Precision in documentation |
Relationship to other roles
Section titled “Relationship to other roles”Context engineers work upstream of AI engineers (they define what context is available) and downstream of domain experts (they encode domain knowledge into retrievable structures). Closely related to platform engineers in large organizations.
Entry paths
Section titled “Entry paths”Data engineer, backend engineer, ML engineer, information architect.
4. AI Engineer
Section titled “4. AI Engineer”Status: Mainstream — the generalist role for building AI-powered products.
What they do
Section titled “What they do”Build end-to-end AI systems. Not researchers (they don’t train models from scratch), but not just integrators either. They take LLMs and orchestration frameworks and build systems that ship. Think of them as software engineers who’ve added LLM integration, evals, and AI product intuition to their stack.
Responsibilities
Section titled “Responsibilities”- Design and implement LLM-powered applications (chatbots, agents, pipelines)
- Build evaluation frameworks to measure model output quality
- Integrate AI capabilities into existing software systems
- Monitor AI systems in production (latency, cost, quality drift)
- Select appropriate models for specific tasks (capability vs. cost tradeoffs)
- Implement fine-tuning or RAG when base models aren’t sufficient
Required skills
Section titled “Required skills”| Technical | Soft |
|---|---|
| Strong software engineering foundations | Product judgment |
| Python (primary), JavaScript (often needed) | Pragmatism over research purity |
| Familiarity with major LLM APIs (Anthropic, OpenAI, Gemini) | Fast iteration mindset |
| Eval design and measurement | Ability to work with ambiguous requirements |
| Understanding of embeddings, RAG, agent frameworks | Communication of AI limitations to stakeholders |
| MLOps basics (deployment, monitoring, versioning) |
The critical distinction from ML Engineer
Section titled “The critical distinction from ML Engineer”AI engineers work with existing models. ML engineers build and train models. In practice, most companies hiring in 2025-2026 need AI engineers (apply the models) not ML engineers (build the models).
Entry paths
Section titled “Entry paths”Software engineer (most common), backend engineer, data engineer, ML engineer transitioning to applied work.
5. LLM Engineer
Section titled “5. LLM Engineer”Status: Specialized variant of AI Engineer, prominent in model-heavy companies.
What they do
Section titled “What they do”Deep specialization in large language model integration and optimization. Where AI engineers are generalists, LLM engineers go deep on the model layer: fine-tuning, RLHF, model selection, prompt optimization at scale, and evaluation infrastructure.
Responsibilities
Section titled “Responsibilities”- Fine-tuning base models for domain-specific tasks
- Designing and running systematic model evaluations (evals)
- Implementing RLHF or similar feedback mechanisms
- Model performance benchmarking and regression testing
- Managing model versions and A/B testing new model releases
- Building tooling for model monitoring and drift detection
Required skills
Section titled “Required skills”| Technical | Soft |
|---|---|
| Python (fluent) | Scientific rigor |
| PyTorch or JAX | Statistical thinking |
| Transformers architecture knowledge | Patience with slow feedback loops |
| Evaluation framework design | Documentation of experiments |
| Distributed training basics |
Where it’s heading
Section titled “Where it’s heading”Strong demand at AI companies (Anthropic, OpenAI, scale-ups) and in large enterprises building proprietary models. Distinct from AI engineer in its proximity to the model itself. Expect this role to bifurcate: pure research at labs vs. applied fine-tuning at enterprises.
6. AI Agent Engineer
Section titled “6. AI Agent Engineer”Status: High growth — one of the most in-demand specialized roles in 2025-2026.
What they do
Section titled “What they do”Design and build autonomous agent systems. While AI engineers build general AI products, agent engineers specialize in systems that plan, reason, use tools, and execute multi-step tasks without constant human intervention.
Responsibilities
Section titled “Responsibilities”- Design multi-agent architectures (orchestrator + specialist agents)
- Build agent memory systems (short-term, long-term, episodic)
- Implement tool use and API integrations for agents
- Design guardrails and safety mechanisms for autonomous systems
- Build human-in-the-loop checkpoints for high-risk decisions
- Monitor agent behavior in production (reliability, cost, anomaly detection)
- Test agent systems systematically (agentic eval is a distinct discipline)
Required skills
Section titled “Required skills”| Technical | Soft |
|---|---|
| Agent frameworks (LangChain, AutoGen, Claude Agent SDK, CrewAI) | Systems thinking |
| Orchestration patterns | Risk judgment (when to let agents act autonomously) |
| Tool/API integration | User experience intuition |
| Async programming | Debugging patience (agents fail in non-deterministic ways) |
| Observability and tracing (LangSmith, Langfuse, etc.) |
Key challenge specific to this role
Section titled “Key challenge specific to this role”Non-determinism. Agent systems fail in ways that are hard to reproduce. Observability tooling (tracing every agent step) is as critical as the agent code itself. Engineers who treat agent debugging like debugging traditional code struggle.
7. Founding AI Engineer
Section titled “7. Founding AI Engineer”Status: Highly sought after in AI-native startups and seed-to-Series A companies.
What they do
Section titled “What they do”A hybrid role unique to early-stage companies: part AI engineer, part product engineer, part technical co-founder. They own core product functionality end-to-end, from architecture decisions to customer interactions, while building on top of AI capabilities.
Typically targets engineers with 0-4 years of experience who are comfortable with ambiguity, figure things out independently, and already use AI tools daily in their workflow.
Responsibilities
Section titled “Responsibilities”- Build entire product features from architecture to deployment, not just assigned tickets
- Make foundational technical decisions that will shape the company’s stack for years
- Work directly with founders on product strategy and prioritization
- Use AI coding tools as force multipliers to ship at startup speed
- Interact directly with early customers to understand problems
- Define engineering culture before it calcifies
What makes this role different
Section titled “What makes this role different”Scope of ownership and ambiguity. A senior engineer at a large company works within defined systems. A founding engineer defines the systems. The leverage is massive in both directions: great decisions compound, bad ones become technical debt that’s hard to escape.
Required profile
Section titled “Required profile”- Bias toward action over analysis paralysis
- Comfort shipping imperfect things and iterating
- Product intuition alongside technical skills
- Already fluent with AI coding tools (Claude Code, Cursor, Copilot)
- Able to context-switch from infra to product to customer research in the same day
Entry paths
Section titled “Entry paths”Strong mid-level engineers at established companies who want more ownership. Common source: engineers who’ve been quietly building side projects with AI tools.
8. AI Architect
Section titled “8. AI Architect”Status: Senior/Staff level — emerging role in larger organizations.
What they do
Section titled “What they do”Design enterprise AI systems at the system level. Where AI engineers ship features, AI architects define the patterns, platforms, and decision frameworks that multiple teams use. They make the technology choices that others live with for years.
Responsibilities
Section titled “Responsibilities”- Define AI technology strategy and stack decisions (which models, which frameworks, which providers)
- Design enterprise AI reference architectures
- Set standards for AI system observability, security, and governance
- Evaluate build vs. buy decisions for AI capabilities
- Ensure AI systems are scalable, cost-effective, and auditable
- Bridge between business requirements and technical AI implementation
Required skills
Section titled “Required skills”- Deep experience across AI/ML stack (models, infrastructure, MLOps)
- Strong communication skills (presenting to C-suite, working with legal/compliance)
- Understanding of cloud provider AI offerings (AWS Bedrock, Azure OpenAI, Vertex AI)
- Security and compliance awareness (GDPR, AI Act, SOC2)
- Experience designing distributed systems at scale
Entry paths
Section titled “Entry paths”Senior AI engineer → Staff → Architect. Often takes 5-8 years in AI-adjacent roles. Alternatively: cloud architect + strong AI self-study.
9. Platform Engineer (AI context)
Section titled “9. Platform Engineer (AI context)”Status: Established role, significantly reshaped by AI.
What they do
Section titled “What they do”Build and maintain the internal developer platform. With AI, this role has expanded to include the “golden path” for AI development: standardized ways for teams to integrate LLMs, common observability infrastructure, cost controls, and guardrails so individual teams don’t reinvent the wheel or create security risks.
AI-specific responsibilities added to traditional platform work
Section titled “AI-specific responsibilities added to traditional platform work”- Provide standardized LLM integration patterns (internal SDKs, proxies, abstractions)
- Manage API keys, rate limits, and cost allocation across teams
- Build AI observability infrastructure (tracing, logging, alerting)
- Enforce security policies for AI outputs (PII filtering, output validation)
- Maintain model registries and versioning systems
- Create “paved roads” for RAG patterns, agent architectures, eval pipelines
Why this role matters more with AI
Section titled “Why this role matters more with AI”When every team is building their own LLM integrations, you get: duplicated cost, inconsistent security, no centralized observability, and no shared learnings. Platform engineers who understand AI prevent this fragmentation. They’re the reason the AI investment in a company scales instead of sprawling.
Required skills (AI additions)
Section titled “Required skills (AI additions)”MLOps tooling, LLM gateway products (LiteLLM, Portkey), cloud AI services, cost optimization patterns, security for AI (prompt injection mitigation, output filtering).
10. Harness Engineer
Section titled “10. Harness Engineer”Status: Emerging — formalized by Martin Fowler in 2025, not yet institutionalized as a standalone title.
What they do
Section titled “What they do”Build the infrastructure that keeps AI agents “under harness” — under control. As agentic AI systems generate code, take actions, and operate with increasing autonomy, harness engineers build the systems that ensure they stay within architectural constraints, produce coherent output, and don’t accumulate entropy over time.
The three pillars
Section titled “The three pillars”1. Context engineering (knowledge infrastructure) Not one-off prompts, but a continuously updated knowledge base embedded in the codebase. Agents know your conventions, architecture decisions, and domain context. Dynamic access to observability data and documentation.
2. Architectural constraints (agent guardrails)
- LLM-based watchdog agents that review generated code
- Custom deterministic linters enforcing your specific architectural patterns
- Structural tests (ArchUnit-style) that run automatically
- Pre-commit hooks that reject code violating established constraints
3. Entropy management (drift prevention) Periodic agents that scan the codebase for: outdated documentation, architectural violations that slipped through, abandoned patterns that reappeared, inconsistencies introduced by multiple agents working in parallel.
The core insight
Section titled “The core insight”Without a harness, AI agents produce code that individually looks fine but collectively drifts away from your architecture, your patterns, and your documentation. The harness is what makes “AI generates most of the code” sustainable at scale rather than a path to unmaintainable systems.
Organizational impact
Section titled “Organizational impact”This role pushes toward intentional technological convergence: organizations with 2-3 primary tech stacks benefit far more from standardized harnesses than organizations with 10 different stacks. It’s a deliberate trade of technical freedom for reliability.
“Ce n’est pas quelque chose dans lequel vous pouvez vous lancer pour des résultats rapides.” — Martin Fowler
Where this role will emerge
Section titled “Where this role will emerge”Currently absorbed by: platform engineers, staff/principal engineers, architecture guilds. Likely to become an explicit role in:
- Companies running autonomous coding agents at scale
- Large enterprises with 50+ engineers using AI coding tools
- Organizations that’ve experienced “AI entropy” firsthand (code that works but nobody understands anymore)
Required skills
Section titled “Required skills”Software architecture, linter/static analysis tooling, LLM orchestration, observability, codebase knowledge management, entropy detection patterns.
11. AI Product Manager
Section titled “11. AI Product Manager”Status: Mainstream and growing, with significant premium over traditional PM roles.
What they do
Section titled “What they do”Product management with deep AI fluency. They understand what AI can and can’t do, manage the unique product challenges of AI systems (non-determinism, latency, hallucinations, cost), and translate between business needs and AI capabilities.
Responsibilities
Section titled “Responsibilities”- Define product requirements for AI features with technical constraints in mind
- Work with AI engineers on evaluation criteria (what does “good” look like?)
- Manage the unique UX challenges of AI: uncertainty, latency, error handling
- Own the cost/quality/speed tradeoffs for AI features
- Communicate AI limitations and risks to stakeholders
- Run A/B tests on model versions, prompt changes, feature changes
What makes AI PM different from traditional PM
Section titled “What makes AI PM different from traditional PM”Traditional PM ships features that behave deterministically. AI PMs ship systems where outputs vary. They need to think probabilistically: not “will this work?” but “what % of the time will this work, and what happens in the other cases?” Quality measurement is continuous, not binary.
Required skills
Section titled “Required skills”Standard PM skills (roadmapping, prioritization, user research) plus: LLM API familiarity, eval design, basic Python for running experiments, understanding of model tradeoffs (accuracy vs. cost vs. latency), AI UX patterns.
Salary context
Section titled “Salary context”FAANG-level: $160K-$200K+ entry-level AI PM. Senior: $200K-$300K+ total compensation.
12. AI Safety & Eval Engineer
Section titled “12. AI Safety & Eval Engineer”Status: Specialized — primarily at AI labs and companies with regulated AI deployments.
What they do
Section titled “What they do”Ensure AI systems behave safely, reliably, and in alignment with intended values. Two related but distinct specializations: Eval Engineers (build systems to measure model behavior) and AI Safety Engineers (identify and mitigate risks in AI systems).
Eval Engineer responsibilities
Section titled “Eval Engineer responsibilities”- Design evaluation frameworks (evals) to measure model quality, safety, and capabilities
- Build automated eval pipelines that run on every model version change
- Define metrics that capture real-world performance (not just benchmark gaming)
- Implement human evaluation workflows for subjective quality dimensions
- Detect regressions before they reach production
AI Safety Engineer responsibilities
Section titled “AI Safety Engineer responsibilities”- Red-team AI systems to find failure modes, jailbreaks, and harmful outputs
- Implement content filtering, output validation, and guardrail systems
- Design human-in-the-loop checkpoints for high-risk decisions
- Monitor production systems for harmful outputs or unexpected behavior
- Work with legal/compliance on AI governance
Required skills
Section titled “Required skills”Rigorous experimental design, statistics, Python, strong understanding of LLM failure modes, communication skills for risk reporting.
Where to find these roles
Section titled “Where to find these roles”Primarily: Anthropic, OpenAI, Google DeepMind, Meta AI, Microsoft AI. Growing in: healthcare, finance, legal tech — regulated industries where AI errors have serious consequences.
13. ML Engineer
Section titled “13. ML Engineer”Status: Established — the most traditional of the AI engineering roles.
What they do
Section titled “What they do”Develop, train, deploy, and maintain machine learning models. In the LLM era, many ML engineers have pivoted toward fine-tuning and applied AI work rather than building models from scratch — that work is increasingly concentrated at a small number of frontier labs.
Responsibilities
Section titled “Responsibilities”- Data pipeline development (collection, cleaning, transformation)
- Model training and fine-tuning
- Feature engineering
- Model serving and deployment (MLOps)
- Performance optimization and model compression
- Production monitoring for model drift
How the role is evolving
Section titled “How the role is evolving”The “build a model from scratch” path is increasingly rare outside frontier labs. ML engineers in most companies now work on: fine-tuning existing models, building RAG systems, deploying and monitoring models in production, and bridging between AI engineers and data infrastructure. The practical overlap with AI engineer is large.
Required skills
Section titled “Required skills”Python (fluent), PyTorch or TensorFlow, distributed computing, data pipeline tools (Spark, Airflow, dbt), cloud ML platforms (SageMaker, Vertex AI, Azure ML), statistical foundations.
14. Career Decision Matrix
Section titled “14. Career Decision Matrix”Which role fits your current background and goals?
| Your current profile | Best next role | Timeline |
|---|---|---|
| Software engineer (3+ years) who wants to work with AI | AI Engineer | 3-6 months upskill |
| Software engineer at early startup who wants ownership | Founding AI Engineer | Now, if opportunity exists |
| Backend engineer interested in infra + AI | Platform Engineer (AI) | 6-12 months |
| Senior engineer who thinks in systems | AI Architect or Harness Engineer | 1-2 years experience accumulation |
| Engineer who likes research and rigor | LLM Engineer or AI Safety/Eval | +ML foundations needed |
| Non-technical who works with AI daily | Prompt Engineer → Context Engineer | 6-18 months |
| PM who wants to stay PM but be more relevant | AI Product Manager | 3-6 months upskill |
| Engineer obsessed with reliability and architecture | Harness Engineer (emerging) | Pioneers’ territory |
The fastest path to AI employment in 2025-2026
Section titled “The fastest path to AI employment in 2025-2026”- Build something with AI APIs (Claude, OpenAI) — a real project, not a tutorial
- Write about what you built (blog post, GitHub README, LinkedIn)
- Add evaluation: measure your system’s quality, show the numbers
- Apply for AI Engineer roles — the bar is demonstrated building, not credentials
Note: 76% of candidates claiming AI expertise lack production-level deployment experience (LangChain State of Agent Engineering 2025). The bar is lower than it appears if you’ve actually shipped something.
15. Salary Benchmarks (2025-2026)
Section titled “15. Salary Benchmarks (2025-2026)”Indicative only — large variance applies. These figures are US market base salaries (2025-2026). Europe runs 30-50% lower, other markets 40-60% lower. Total compensation (equity, bonus, RSUs) can significantly exceed base, especially at startups and FAANG. Experience level, location within a country, company stage, and negotiation all create wide variance. Use these as orientation, not negotiation anchors.
| Role | Entry | Mid | Senior | Notes |
|---|---|---|---|---|
| Prompt Engineer | $80K-$110K | $110K-$150K | $150K-$180K | Shrinking standalone market |
| Context Engineer | $100K-$140K | $140K-$180K | $180K-$230K | Growing fast |
| AI Engineer | $120K-$160K | $160K-$220K | $220K-$300K | Highest volume of open roles |
| LLM Engineer | $130K-$170K | $170K-$250K | $250K-$350K | Lab-level roles higher |
| AI Agent Engineer | $130K-$170K | $170K-$240K | $240K-$320K | Strong demand 2025-2026 |
| Founding AI Engineer | $100K-$150K + equity | — | — | Equity makes total comp wide-ranging |
| AI Architect | — | $180K-$260K | $260K-$380K | Senior/Staff only |
| Platform Engineer (AI) | $110K-$150K | $150K-$210K | $210K-$280K | |
| Harness Engineer | Not yet standardized | — | — | Absorbed into other roles |
| AI Product Manager | $130K-$170K | $170K-$230K | $230K-$350K | FAANG premium significant |
| AI Safety/Eval Engineer | $140K-$180K | $180K-$250K | $250K-$400K | Lab compensation highest |
| ML Engineer | $100K-$140K | $140K-$200K | $200K-$280K | Lower demand outside labs |
Sources: FinalRoundAI (2025), Alcor AI Salary Report (2025), RiseWorks AI Talent Report (2025), job postings analysis.
16. What’s Not a Role (Yet)
Section titled “16. What’s Not a Role (Yet)”Some terms you’ll hear that describe practices or methodologies, not job titles:
Vibe coder — A methodology (use AI coding assistants to handle implementation while you focus on design), not a job. Andrej Karpathy coined the term then himself pivoted toward “context engineering” as more precise. No serious company has “Vibe Coder” on a job description.
AI-native engineer — Describes a quality expected of all engineers increasingly, not a specialized role. It means: you use AI tools fluently in your daily workflow. It’s the bar, not the title.
Orchestration engineer — Sometimes used for agent systems, overlaps significantly with AI Agent Engineer. Not yet a distinct category.
These terms are worth knowing (you’ll encounter them in job descriptions and articles) but don’t represent distinct career paths — yet.
17. Job Listings
Section titled “17. Job Listings”Coming soon — Curated listings for AI roles at companies building seriously with Claude Code and agentic AI.
If you’re hiring for any of the roles described in this guide, reach out to discuss featuring your opportunity here.
See Also
Section titled “See Also”- Learning to Code with AI — skill development for developers using AI
- AI Ecosystem: Tools & Integrations — which tools each role uses
- Methodologies — TDD, SDD, BDD workflows relevant to AI engineers
- Architecture — how Claude Code works, relevant for AI agent engineers
- Security Hardening — critical reading for AI Safety engineers and Platform engineers