Skip to content

How Roles Are Changing

The transformation described in the chapter opening—Scale AI growing from $250 million to $2 billion—wasn't just revenue growth. It was workforce evolution. By 2024, 12% of Scale's contributors held PhDs and over 40% had advanced degrees1. The data labelers who became RLHF specialists didn't just change titles. They developed genuinely new expertise.

That pattern—not replacement, but evolution—defines what's happening across every function in AI-first organizations.

Three Emerging Roles

When Anthropic posted a job listing for "Prompt Engineer and Librarian" with compensation up to $335,000, it signaled something profound: AI-first companies were creating roles that didn't exist three years ago2. Three positions have emerged as distinctive—and all three are simultaneously essential and temporary.

Prompt engineers evolved from "person who writes good ChatGPT prompts" to a systematic discipline in about 18 months. At production-critical companies, they don't just write prompts—they build infrastructure: versioned prompt libraries, A/B testing frameworks, safety measures against prompt injection. The salary range reflects this technical depth: entry-level earns $85,000-$140,000, with senior roles commanding $220,000-$350,0003. Richard Socher, former Salesforce Chief Scientist, argues that prompt engineering should be viewed "in terms of delegation"—a skill that "the most productive people and best performers should have"4.

Agent operators emerged when AI systems started doing multi-step work without human intervention. Someone needs to watch what happens when you tell an AI to "book the cheapest flight to Tokyo" and it starts navigating websites, filling forms, and making autonomous decisions. OpenAI's launch of Operator in January 2025 validated this role's importance—CEO Sam Altman called autonomous agents "the next giant breakthrough beyond incremental model improvements"5. The work combines system administration skills with domain judgment—knowing when an agent's "smart" solution is actually the kind of dangerous idea a talented junior engineer might propose.

AI product managers bridge technical capabilities with business value in fundamentally different ways than traditional PMs. The shift: from scoping features to scoping capabilities, from deterministic outcomes to probabilistic ones. Traditional PRD: "Add export functionality with CSV and PDF options." AI-era PRD: "Build LLM-based export that adapts format based on user context, with 85% accuracy threshold and fallback to manual selection." Compensation reflects this complexity—AI PMs at leading companies earn $192,000 to $456,0006.

Four Transforming Roles

While new roles emerge, existing roles transform. Engineers, product managers, designers, QA—none disappear. All change profoundly.

flowchart LR
    subgraph BEFORE["<b>Traditional Focus</b>"]
        direction TB
        ENG1[Engineer:<br/>60% coding, 40% architecture]
        PM1[PM:<br/>Deterministic specs]
        DES1[Designer:<br/>Individual screens]
        QA1[QA:<br/>Expected = Actual]
    end

    subgraph AFTER["<b>AI-Augmented Focus</b>"]
        direction TB
        ENG2[Engineer:<br/>70% architecture, 30% coding]
        PM2[PM:<br/>Probabilistic outcomes]
        DES2[Designer:<br/>System constraints]
        QA2[QA:<br/>Relevance + Safety + Accuracy]
    end

    ENG1 -->|"Inversion"| ENG2
    PM1 -->|"Shift"| PM2
    DES1 -->|"Expand"| DES2
    QA1 -->|"Redefine"| QA2

    style AFTER fill:#1a8a52,stroke:#14693e

Software engineers now spend 70% of time on architecture, system design, and validating AI outputs—only 30% on direct coding7. Code review focuses on different failure modes than human code: hallucination detection, architectural coherence, security vulnerabilities where AI still makes mistakes. According to Qodo's 2025 research, 81% of teams using AI for code review reported quality improvements—but only 3.8% report high confidence shipping AI code without human review8. That last statistic matters. AI amplifies engineering judgment, not replaces it.

But the gains aren't straightforward. A METR study found experienced developers took 19% longer on mature open-source projects when using AI tools—despite feeling 20% faster9. The illusion of speed is real, particularly for senior engineers working on complex, established codebases.

Product managers now write specifications for probabilistic systems. Two years ago: 15-page PRDs taking weeks. Now: AI generates initial requirements from user feedback analysis in hours, PMs focus on strategic decisions and tradeoff analysis. Communicating uncertainty becomes a core skill—defining confidence thresholds, graceful degradation, and success metrics beyond binary pass/fail. By year-end 2025, over 60% of enterprise software teams shipped at least one LLM-powered feature10. Every PM working on those teams learned to manage probabilistic outcomes—or failed.

Designers spend less time creating individual assets and more time defining systems that accommodate AI-generated variations. The shift: from pixel-perfect screens to component behaviors and relationships. Figma AI plugins generate components matching existing design systems and convert designs directly into functional code11. Designers define the constraints; AI fills in the variations.

QA engineers face the deepest transformation. Traditional testing relies on assertions: expected equals actual. LLM outputs vary with each interaction, breaking that fundamental model. QA now assesses responses on multiple criteria—relevance, safety, accuracy, brand alignment—all against thresholds rather than exact matches. By 2025, 75% of organizations identified AI-driven testing as pivotal, though only 16% had actually adopted it12. The gap creates competitive advantage.

The Meta-Pattern: Specialist Today, Universal Tomorrow

flowchart TB
    subgraph TODAY["<b>Today: Specialist Roles</b>"]
        direction LR
        PE[Prompt Engineer<br/>$85K-$350K]
        AO[Agent Operator<br/>Monitors autonomous AI]
        APM[AI Product Manager<br/>Probabilistic specs]
    end

    subgraph TOMORROW["<b>Tomorrow: Distributed Skills</b>"]
        direction LR
        ALL[Every Role]
        S1[Prompt engineering<br/>becomes universal]
        S2[Agent oversight<br/>becomes domain work]
        S3[AI judgment<br/>becomes standard PM]
    end

    PE -->|"Skills distribute"| S1
    AO -->|"Skills distribute"| S2
    APM -->|"Skills distribute"| S3

    S1 --> ALL
    S2 --> ALL
    S3 --> ALL

    style TODAY fill:#c77d0a,stroke:#a06508
    style TOMORROW fill:#1a8a52,stroke:#14693e
    style ALL fill:#1e6fa5,stroke:#155a85,stroke-width:2px

Job postings requiring prompt engineering skills increased 434% from 2023 to 202413. But that same data shows the skills spreading beyond specialists. Non-technical professionals across finance, marketing, legal, and operations now develop "AI fluency" as a core competency14. No-code AI tools achieved 72% adoption among startups in 2024, reducing coding time by 45%15.

Elena Verna, Head of Growth at Loveable, observed the cultural shift: "AI-native employees simply default to solving everything with AI. It's embedded in every part of their workday"16. When AI operation becomes reflexive rather than specialized, the dedicated operator role converges back into domain roles.

The pattern: Every "emerging role" teaches the organization what everyone eventually needs to learn. Prompt engineering becomes a universal skill. Agent operation becomes part of domain expertise. AI product judgment becomes standard product work.

Scale AI improved labeling accuracy by 35% over competitors through technology investments and workforce development17. The lesson: transformation isn't just about tools. It's about people learning to use them.

What matters isn't whether to hire these specialists—it's how long before these skills distribute across your entire organization, and what you do to accelerate that convergence rather than delay it.


References


← Previous: The 90-Day AI Fluency Program | Chapter Overview | Next: Building AI Culture →