Advice to pivot into AI Safety is uncalibrated and thus suspect

Aligning our advice about aligning AI

2025-09-28 — 2025-10-05

Wherein the AI‑safety career‑advice ecosystem is described and its costly lack of mooring to ground-truth or optimality is noted.

AI safety
catastrophe
economics
faster pussycat
innovation
machine learning
incentive mechanisms
institutions
networks
wonk

Assumed audience:

Mid career technical researchers considering moving into AI Safety research, career advisors in the EA/AI Safety space, AI Safety employers and grantmakers

Nonetl;dr

AI career advice orgs, prominently 80,000 Hours, encourage career moves into AI risk roles, including mid‑career pivots into roles in AI safety research labs.

Without side information, that advice is not credible for mid‑career readers, because it does not have a calibration mechanism to ensure that the expected value of attempting a pivot is positive. Advice organizations influence beliefs and enlarge funnels but don’t bear most costs when beliefs overshoot, or have informative feedback channels about acceptance rates. As such, the system predictably dissipates value for some applicants and for the field.

The analysis here is should not be too surprising; we all know advice calibration is hard, and it is easy to deduce that this advice isn’t. I’m sure other people have modeled it. Nonetheless, I couldn’t find a reference tying it all up into a bow and giving decision rules, so I worked it out myself. Spelling it out again seems worthy since the costs of leaving it unaddressed are high.

Solutions, both personal and institutional, are proposed below.

NoteMeet Alice

Alice is a senior software engineer in her mid-30s. She has been donating roughly 10% of her income to effective charities and now wonders whether to switch lanes entirely. She has saved six months of runway to explore AI-safety roles — research engineering, governance, or technical coordination — but each month out of work costs her real money and momentum. Her question is simple: Is this pivot worth the gamble?

Figure 1: The totem pole of AI safety careers. Atop the column is the muse of alignment, inspiring the career advice orgs that amplify interest in AI safety roles. At the base sit recruiters, who screen and filter applicants. In the corner, the lion of ill‑advised unemployment devours those who miscalibrate their EV.

Here’s a sketch of a toy model of the AI‑safety career‑advice economy as it stands, with implications for technical researchers considering a pivot into AI‑safety work—especially mid‑career researchers or people who will pay an opportunity cost to pivot. I argue the current pipeline, in expectation, imposes a Pivot Tax on mid‑career candidates and that this is predictable from first principles.

This is a practical issue, but a warning sign of a structural problem: If the field that studies alignment relies on misaligned mechanisms, it weakens our credibility because we fail to practise what we preach.

For the purposes of this note, an AI pivot means ‘applying to roles in existing AI safety organizations’ (labs, fellowships, think tanks, etc.). Other paths exist—intrapreneurship inside one’s current org, founding a new team, building tools or consultancies—and sometimes advice orgs do recommend these. We return later to how such alternatives may change the game.

This logic likely extends to other impact fields with constrained hiring pipelines, e.g., climate tech, biosecurity, global health, and so on, but I know less about those.

In this model we have a recruiting process with weak control mechanisms and marginal expected value (EV). Advice orgs expand application flow but do not internalize the applicant‑side downside, so they eventually give advice that benefits neither the candidate nor the field.

The institutional alignment problem is that the visible proxy — growth in AI safety applications — is cheap to count, while the target — welfare‑maximizing matches to roles at low opportunity cost — is expensive and under‑observed. Optimizing the proxy without a feedback loop predictably induces miscalibration: beliefs overshoot and attempts exceed the optimum, costing both individuals and the field.

Employers rarely publish base rates, i.e. “how many people applied to and passed through each stage of the hiring funnel”. Advice orgs, AFAICS, never publish advisor‑forecast calibration. So the advice ecosystem is missing key feedback loops that would allow it to self‑correct. Without those, it’s rational to treat the generic encouragement to “do AI safety stuff” as over‑optimistic.

TipKey terms
  • Advice organisations — groups that recommend career moves and shape beliefs (e.g., career-advice nonprofits).
  • Employers — organisations that offer and fill paid roles.
  • Impact — value created, whether directly through work or indirectly via donations, both measured in donation-equivalent dollars.
  • Private costs — the candidate’s own losses: sabbatical spend, lost wages, opportunity cost.
  • Public or social costs — losses at the field level when miscalibrated entry burns collective value.
  • Pivot tax — shorthand for the predictable loss that occurs when candidates over-enter a narrow funnel.

The so-called “Pivot tax” is not a literal tax (i.e. transfer) but deadweight loss: surplus destroyed when miscalibrated entry dissipates value rather than reallocating it. But Pivot Dissipation has too many syllables.

I’m mostly saying “advice organisations” and “employers”, rather naming specific organisations such as “80,000 Hours” or “FAR.ai” or whatever. They are in fact the main player in this space. Nonetheless, I want to keep the discussion about general organisations rather than naming specific ones because

  1. The analysis applies to any advice org or employer with similar incentives, and similar information asymmetries, and
  2. as a way of distancing myself from any moral judgment of specific organisations.

I see this very much as a systems-level problem, rather than a criticism of specific actors.

1 Part A—Private decision model: Alice’s gamble

An uncertain career pivot is a gamble, and as such we model it the usual way that we model gambles.

Suppose you have a stable job, and make donations to your impact causes, but you think you might be able to pivot into something more closely aligned to AI and thus achieve higher impact. You take an unpaid sabbatical to prepare and apply for a pivot into an AI safety role, maybe at a lower wage but with higher impact.

We assume that you have preference both for remuneration and for impact, and a valuation of potential impact in the new role; later on we can solve for potentially-different valuations by employers. The scope here emphasises mid‑career professional considering an unpaid sabbatical to pivot into an AI‑safety role. We evaluate the decision in donation‑equivalent dollars (after‑tax wage + donations you would actually make + your impact valuation). We discount future benefits and costs at a continuous rate \(\rho>0\) (per year). All flows below are valued in present value (PV) donation‑equivalent post-tax dollars. We use a very simple linear utility function, where you value \(\$1\) of personal consumption like \(\$\alpha\) of impact/donations.1 We don’t deal with risk aversion here, or diminishing returns to keep things simple. We also ignore job satisfaction and other non-pecuniary benefits, but it is easy to add those in, or count them as part of the wage. Let

  • \(u=w+\alpha(i+d)\) be annual “utility in donation‑equivalent dollars”, with \(\alpha\ge 0\) your weight on (impact + donations) relative to personal consumption.
  • Your baseline annual utility if you stay put is \(u_0:=w_0+\alpha(i_0+d_0)\), i.e. your wage \(w_0\), your impact \(i_0\) (maybe zero), and your donations \(d_0\) in your current role.
  • Your annual utility in the new role if the pivot succeeds will be \(u_1:=w_1+\alpha(i_1+d_1)\), i.e. new wage, new donation rate etc.
  • Define \(\Delta u := u_1 - u_0\) (per-year surplus in donation-equivalent dollars if the pivot succeeds).
  • The per‑application success probability is \(p\) (i.i.d. across applications in this toy model).
  • Job opportunities arrive according to a Poisson process with rate \(r\) (opportunities/year).
  • The maximum sabbatical length (the “runway”) is \(\ell\) (in years).
  • We discount future benefits and costs at rate \(\rho>0\) (per year). If you prefer to think in terms of a planning horizon, you can approximate \(\rho \approx 1/T\) where \(T\) is your “effective horizon” in years, say your AGI timeline.
  • If you get a new job at time \(\tau\), you start immediately and receive the surplus \(\Delta u\) per year from then on (discounted at rate \(\rho\)).
  • If you burn your runway \(\ell\) without success, you return to your old job (which feels optimistic in this economy, so beware).
  • While on sabbatical you burn value at rate \(c\) (k$/year, donation-equivalent). The PV over \([0,\min\{T_1,\ell\}]\) is \(\int_0^{\min\{T_1,\ell\}} c\,e^{-\rho t}\,dt\). If you fail (spending the full \(\ell\)), the PV is \(C_{\text{fail},\rho}=c\,\frac{1-e^{-\rho\ell}}{\rho}\). Here \(c\) captures all net opportunity costs during sabbatical not already counted in \(\Delta u\) (foregone pay, benefits, lost progression, living costs, etc.).

Success in the sabbatical. If our application rate is a Poisson process with rate \(r\), then our success rate is also a Poisson process with rate \(p r\). More specifically, it’s a stopped Poisson process, because we stop either after our first success or if more than \(\ell\) time passes. The duration of the sabbatical \(\tau\) is the minimum of the time to first success, \(\tau_1 \sim\mathrm{Exp}(rp)\), and the maximum sabbatical length, \[ \tau \approx \min\{\tau_1,\ell\}. \] Since \(P(\tau_1>\ell)=e^{-rp\ell}\), the expected sabbatical length is (e.g. by the survival function method), \[ \begin{aligned} \mathbb{E}[\tau]&= \frac{1-e^{-r p \ell}}{r p}\\ \mathbb{E}[\tau\mid \text{success}]&=\dfrac{1}{r p}-\dfrac{\ell e^{-r p \ell}}{1-e^{-r p \ell}} \end{aligned} \] And the success probability is: \[ q = P(\tau_1 \le \ell) = 1-e^{-r p \ell}. \]

You only get \(\Delta u\) if you land the pivot job before the runway ends. Let \(\lambda:=r p\) be the success hazard while on sabbatical and \(T_1\sim\mathrm{Exp}(\lambda)\) the time to first success. The sabbatical ends at \(\tau=\min\{T_1,\ell\}\).

With discounting at rate \(\rho\), the present value of attempting the pivot is \[ \boxed{ \Delta \mathrm{EV}_\rho(p) =\frac{1-e^{-(\lambda+\rho)\ell}}{\lambda+\rho}\,\Big(\frac{\Delta u\,\lambda}{\rho}-c\Big), \qquad \lambda=r p.} \] Intuition: the prefactor \(\frac{1-e^{-(\lambda+\rho)\ell}}{\lambda+\rho}\) is the discounted expected length of the sabbatical “clock”, and the bracket compares the discounted gain rate \(\Delta u\lambda/\rho\) to the burn rate \(c\).

Sanity check: if \(p=0\), then \(\lambda=0\) and \(\Delta \mathrm{EV}_\rho(0)=-c\frac{1-e^{-\rho\ell}}{\rho}=-C_{\text{fail},\rho}\) (you only burn runway). As \(p\to1\), \(\Delta \mathrm{EV}_\rho(p)\to \Delta u/\rho\) (immediate success, no exploration cost in the limit).

1.1 Break-even \(p_\rho^*\) (closed form)

The sign of \(\Delta \mathrm{EV}_\rho(p)\) is determined by the bracket. Thus the break-even per-application success probability is \[ \boxed{\,p_\rho^*=\frac{c\,\rho}{r\,\Delta u}\,}. \] Comparative statics (all else equal):

  • Higher burn \(c\) or higher discount rate \(\rho\) \(\Rightarrow\) increase \(p_\rho^*\) (making it harder for us to break even).
  • Higher shot rate \(r\) or higher upside \(\Delta u\) \(\Rightarrow\) decrease \(p_\rho^*\).
  • \(p_\rho^*\) is independent of runway \(\ell\) (though \(\ell\) still scales the level of \(\Delta \mathrm{EV}_\rho\)).

A tiny-probability expansion for intuition. When \((r p+\rho)\ell\ll1\), \[ \Delta \mathrm{EV}_\rho(p)\approx \ell\big(\Delta u\,r p/\rho - c\big), \] So, the same closed-form threshold drops out immediately.

1.2 Worked example

Cool! Now, let’s plug in some plausible representative numbers for Alice, our mid‑career technical researcher considering a pivot into an AI safety role in a developed economy. All dollar values are in thousands of USD.

  • Personal+impact weights. Set \(\alpha=1\) (we value \(\$1\) of impact/donations like \(\$1\) of personal spend).
  • Baseline. \(w_0=180\), \(d_0=18\), \(i_0=0\)
  • Target role. \(w_1=120\), \(d_1=0\), \(i_1=100\)
  • Process. \(\ell=\tfrac12\) years (6 months), \(r=24\) apps/year. [TODO clarify]
  • Discount & runway costs. \(\rho = 1/3 \approx 0.333\,\mathrm {yr}^{-1}\) (roughly “3-year horizon” feel), \(c=25\).

This gives us \(\Delta u = (120 + 0 + 100) - (180 + 18 + 0) = 22\).

Then the break-even probability is \[ p_\rho^*=\frac{c\rho}{r\Delta u} =\frac{25\cdot(1/3)}{24\cdot 22} \approx \boxed{1.58\%}\ \text{per application.} \] Over a 6-month runway, there’s a \(q_\rho^* = 1-e^{-r p_\rho^* \ell} =1-e^{-12\cdot 0.01579} \approx \boxed{17.3\%}\) chance of at least one success.

At \(p_\rho^*\), her truncated expected sabbatical is \[ \mathbb{E}[\tau]=\frac{1-e^{-r p_\rho^* \ell}}{rp_\rho^*}\approx 0.455\,\text{yrs}\,(\approx 5.5\ \text{months}), \quad \mathbb{E}[\tau\mid\text{success}]\approx 0.240\,\text{yrs}\,(\approx 2.9\ \text{months}), \] This matches our intuition. Choosing \(\rho\approx 1/3\) is a steep discount, but seems to be pretty common in people who are worried about short-timeline AI.

Let’s plot a few values to visualize the trade-offs.

Code
# Discounted EV of an AI-safety pivot (PRIVATE MODEL)
# Replaces the T-horizon version with continuous-time discounting at rate rho.

import numpy as np
import matplotlib.pyplot as plt

# Optional styling: keep portable
try:
    from livingthing.matplotlib_style import set_livingthing_style
    set_livingthing_style()
except Exception:
    pass

# -----------------------------
# Model primitives (edit these)
# -----------------------------
# Process
r    = 24.0      # application opportunities per year
ell  = 0.5       # runway (years)
rho  = 1.0/3.0   # discount rate per year (~"3-year horizon" feel)
c    = 25.0      # runway burn rate (k$/year, donation-equivalent)

# Upside options (k$/year): Δu = (w1 + i1 + d1) - (w0 + i0 + d0) with α=1 already folded in
delta_u_list = [5.0, 22.0, 42.0]  # try a few values to see sensitivity

# -----------------------------
# EV function (exact, closed-form, discounted)
# -----------------------------
def ev_discounted(p, delta_u, r, ell, rho, c):
    """
    Discounted expected value ΔEV_ρ (in k$ PV) of attempting a pivot with max runway ell.
    Formula: ΔEV_ρ(p) = [(1 - exp(-(r*p + rho)*ell)) / (r*p + rho)] * (delta_u*(r*p)/rho - c)
    Handles vector p; rho>0.
    """
    p = np.asarray(p, dtype=float)
    lam = r * p
    S = lam + rho
    scale = (1.0 - np.exp(-S * ell)) / S
    return scale * (delta_u * lam / rho - c)

def p_star_discounted(r, rho, c, delta_u):
    """
    Closed-form break-even p*: p* = (c * rho) / (r * delta_u).
    Returns np.nan if delta_u<=0 (no upside).
    """
    return np.nan if delta_u <= 0 else (c * rho) / (r * delta_u)

def q_over_runway(p, r, ell):
    """Total success probability over runway ell when arrivals ~ Poisson with rate r*p."""
    return 1.0 - np.exp(-r * p * ell)

# -----------------------------
# Plot EV vs p (log x-axis), annotate p*
# -----------------------------
p_grid = np.logspace(-3.0, 0, 1000)  # p from 0.1% to 100%

plt.figure(figsize=(8, 5.5))
for du in delta_u_list:
    ev_vals = ev_discounted(p_grid, du, r, ell, rho, c)
    p_star = p_star_discounted(r, rho, c, du)
    label = f"Δu={du:.0f}k/y"
    if np.isfinite(p_star) and (p_grid.min() <= p_star <= p_grid.max()):
        q_star = q_over_runway(p_star, r, ell)
        label += f"  (p*≈{100*p_star:.2f}%,  q*≈{100*q_star:.1f}%)"
        # Mark the root
        y_star = ev_discounted(p_star, du, r, ell, rho, c)
        plt.plot([p_star], [y_star], "o")
    plt.plot(p_grid, ev_vals, label=label)

plt.axhline(0.0, color="gray", linestyle="--", linewidth=1)
plt.xscale("log")
plt.xlabel("Per-application success probability p (log scale)")
plt.ylabel("ΔEV (present value, k$ donation-equivalent)")
plt.title(f"Discounted EV vs p   (r={r}/y, ell={ell}y, rho={rho:.3f}/y, c={c}k/y)")
plt.legend()
plt.grid(True, which="both", ls=":", alpha=0.6)
plt.tight_layout()
plt.show()

# -----------------------------
# Numeric table for the worked example Δu=22k/y
# -----------------------------
du = 22.0

# Baselines: p=0 (pure burn) and p→1 (immediate success)
ev_p0 = ev_discounted(0.0, du, r, ell, rho, c)
ev_p1_limit = du / rho  # as p→1, ΔEV → Δu/ρ
print(f"Baseline (p=0, pure burn): ΔEV = {ev_p0:.2f} k$ (PV)")
print(f"Upper bound (p→1, immediate success): ΔEV → {ev_p1_limit:.2f} k$ (PV)")
print()

for pct in [0.01, 0.02, 0.03]:
    val = ev_discounted(pct, du, r, ell, rho, c)
    print(f"ΔEV(Δu=22k/y, p={100*pct:.0f}%): {val:+.2f} k$ (PV)")

p_star = p_star_discounted(r, rho, c, du)
if np.isfinite(p_star):
    q_star = q_over_runway(p_star, r, ell)
    print(f"Break-even p* (Δu=22k/y): {100*p_star:.3f}% per application")
    print(f"q* over runway (ell={ell}y): {100*q_star:.2f}%")
else:
    print("No break-even (Δu<=0).")

Baseline (p=0, pure burn): ΔEV = -11.51 k$ (PV)
Upper bound (p→1, immediate success): ΔEV → 66.00 k$ (PV)

ΔEV(Δu=22k/y, p=1%): -3.98 k$ (PV)
ΔEV(Δu=22k/y, p=2%): +2.74 k$ (PV)
ΔEV(Δu=22k/y, p=3%): +8.75 k$ (PV)
Break-even p* (Δu=22k/y): 1.578% per application
q* over runway (ell=0.5y): 17.25%

If you wish to play around with the assumptions check out the interactive Pivot EV Calculator (source at danmackinlay/career_pivot_calculator).

For Alice, the threshold probability \(p_\star\) marks the point where a pivot breaks even; below that, she burns runway faster than expected value, and if she wants to make impact, she might consider donating money instead.

If your inferred success probability \(p\) is below your personal break-even \(p_\star\), delay the pivot. Seek early, cheap signals of fit before burning runway.

1.3 Are we applying too much as individuals?

I reckon so. Maybe we’ve already passed the break‑even point—maybe not.

My reasoning is that although the break‑even \(p_\rho^*\) is low (≈1.6% per application with \(\rho\approx 1/3\,\mathrm{yr}^{-1}\) in the worked example), this is a generous lower bound, and I expect the actual break‑evens to be higher. The threshold is also sensitive to \(\rho\) and \(\Delta u\).

For one thing, the calculations are very sensitive to the upside \(\Delta u\); if your impact valuation is lower, or your wage cut is larger, then \(p_\rho^*\) rises quickly. If you’re simply “wrong” about your impact valuation, you can quickly get negative EV.

Let’s be real: we have no idea what the impact of any given career pivot will be. Perhaps it’s a black swan farming situation where expected value isn’t even a productive framing?

By the same token, opportunity costs are still real. If you’re prepared to pay a substantial pivot cost to achieve a marginal impact, you might instead donate that pivot cost to your favourite AI safety NGO and end up ahead with no risky gambles involved.

For one thing, job applications are not IID; rather, they’re correlated because we apply to similar roles with similar CVs. The effective number of independent shots is lower than the raw number of applications.

And what even is \(p\)? Trying to work out that one was what got me started running this calculation for myself. That is one thing we don’t get from the career advice. Without that, many mid‑career readers will implicitly assume \(p>p_\star\) when in fact \(p<p_\star\).

The remedy isn’t complicated in principle: publish base rates. Employers can publish stage counts; advice organizations can publish calibrations of their advisors’ forecasted \(p\) against realized outcomes. Absent that, the rational prior for a mid‑career person is that generic encouragement to “do AI safety stuff” is EV‑negative unless we have private side information that pushes our \(p\) above our personal \(p_\star\).

2 Part B — Field-level model: Oversubscription and welfare

I just realised that I conflated private utility and public good in this part. I should fix that. But not today, I need to sleep. Anyway, it won’t change the main conclusions, which are still directionally correct, just shift the breakeven points.

So far we’ve been talking about the pivot with respect to a candidate’s private preferences and costs.

Let’s generalize from you and me and Alice to the field, and try to model when another applicant helps at the margin.

For this to be anything but trivial, we’ll have to include some way of capturing heterogeneity in candidate quality, otherwise everyone is identical and only one person should ever apply for any job.

We add the minimal amount of heterogeneity to make the problem interesting.

2.1 Model

  • Each applicant \(k\in{1,\dots,K}\) has a true per‑year impact in the role \(I^{(k)}\sim F\), i.i.d. (measured in donation-equivalent \(\Delta u\)-units per year, matching the Part 1 scale).

  • Employers observe \(I^{(k)}\) (or a sufficiently informative proxy) and hire the top \(N\). Applicants do not observe their own \(I^{(k)}\). We treat private candidate losses as independent of \(I^{(k)}\): each failed pivot costs \(c \ell\) of value regardless of talent.

  • Let the field‑side annual impact of the hires be \[ S_{N,K}:=I_{(K)}+I_{(K-1)}+\dots+I_{(K-N+1)}. \] We care about \(B(K):=\mathbb{E}[S_{N,K}]\), and in particular its marginal value for one extra applicant. \[ \mathrm{MV}_K:=B(K+1)-B(K). \]

  • Each unsuccessful applicant pays the discounted sabbatical-cost PV \[ \boxed{\,C_{\text{fail},\rho}=c\int_0^\ell e^{-\rho t}\,dt=c\,\frac{1-e^{-\rho\ell}}{\rho}\,}. \] Under the sabbatical-then-stop rule, all \((K-N)\) unsuccessful candidates pay the full sabbatical cost.

We do not model org‑side congestion (it’s empirically small because employers stop looking at candidates when they’re overwhelmed (J. Horton and Vasserman 2021)). All the benefit in the model comes from the fact that, all else being equal, more applicants mean a better top‑\(N\).

  1. With \(N\) seats fixed, adding one more applicant increases the expected number of failures by exactly 1. Hence the marginal private cost of one extra applicant is \(C_{\text{fail},\rho}\), the expected sabbatical burn per failed candidate.

  2. Whether adding applicants is socially worthwhile depends on \(\mathrm{MV}_K\), i.e., on the right hand tail of \(F\).

An interesting question is: how much better does one candidate make the top-\(N\)? For sufficiently many candidates \(K\), this depends only on the tail of the candidate quality distribution, \(F\), because of extreme value theory (which gives us a kind of “law of large numbers for maxima”). Now, what kind of extreme value distribution does \(F\) have? The default would be a light-tailed distribution such as Exponential or Normal, but it’s also plausible that \(F\) has a heavy tail, e.g., a Pareto or Fréchet distribution. Picking a heavy-tailed distribution is a way of capturing the intuition that there are “unicorn” candidates (e.g., 10x engineers) who are much better than the rest. I think the candidate quality distribution is heavy-tailed, but how heavy-tailed it is matters a lot, and I don’t know.

Anyway, let’s introduce some families of distributions and see how \(\mathrm{MV}_K\) behaves.

2.1.1 Light tails

The Exponential distribution is nice because it is exact. If \(I\sim\text{Exp}(\lambda)\) on \([0,\infty)\), the exact identity is \[ B(K)=\frac{N}{\lambda}\Bigl(1+H_K-H_N\Bigr). \] Therefore \[ \boxed{\ \mathrm{MV}_K=B(K+1)-B(K)=\frac{N}{\lambda}\cdot\frac{1}{K+1}\ }. \] Diminishing returns are hyperbolic. Each additional applicant contributes \(\tfrac{N}{\lambda(K+1)}\) expected impact‑units per year.

2.1.2 Heavy tails

Fréchet distribution is heavy-tailed and not too bad to work with maximum statistics.

If \(I\sim\text{Fréchet}(\alpha,s)\) with \(\alpha>1\) (finite mean), \[ B(K)=s\,K^{1/\alpha}\,\underbrace{\sum_{k=1}^{N}\frac{\Gamma\bigl(k-\tfrac{1}{\alpha}\bigr)}{\Gamma(k)}}_{=:C_N}+o\bigl(K^{1/\alpha}\bigr). \] Therefore, for large \(K\), \[ \boxed{\ \mathrm{MV}_K\approx \frac{s\,C_N}{\alpha}\,K^{\frac{1}{\alpha}-1}\ }. \] Diminishing returns are much slower: \(\mathrm{MV}_K\propto K^{-(1-1/\alpha)}\).

  • If \(\alpha=2\): \(\mathrm{MV}_K\sim K^{-1/2}\) and there are some exceptionally strong candidates who you need to audition a lot of people to fin
  • As \(\alpha\downarrow 1\): \(\mathrm{MV}_K\) decays extremely slowly, and the field is full of unrecognised supergeniuses.

2.2 Oversubscription threshold

When does persuading one more applicant start to reduce total welfare?

Every extra applicant slightly improves the chance that the best hires are stronger, but adds one more failed attempt that burns private resources. We compare these in present-value dollars, using the same continuous discount rate \(\rho\) we introduced earlier. This single rate reflects both time preference and any expected shortening of future impact streams (e.g. job changes, role turnover, project endings).

A filled role produces a stream of impact over time. When we discount that stream at rate \(\rho\), the total present value of \(\$1K\)/year of annual impact is \(1K/\rho\). This already accounts for the fact that future years are uncertain and less valuable in expectation.

\[ \boxed{H_\rho=\int_0^\infty e^{-\rho t}\,dt=\frac{1}{\rho}} \qquad\text{(PV weight of one unit of per-year impact).} \]

The planner’s break-even condition for admitting one more applicant is \[ \boxed{\,\mathrm{MV}_K\cdot H_\rho \;\ge\; C_{\text{fail},\rho}=c\,\frac{1-e^{-\rho\ell}}{\rho}\,}. \]

Adding more applicants helps the field only while the discounted value of a slightly better hire (the left side) exceeds the discounted expected cost of one more failed pivot (the right side).

Solving for the critical pool size \(K_\rho^\dagger\) (beyond which the marginal extra applicant has a net negative effect) gives:

  • Exponential (\(I\sim\mathrm{Exp}(\lambda)\), mean \(1/\lambda\)): \[ \boxed{\,K_\rho^\dagger = \frac{N}{\lambda}\cdot\frac{1}{\rho\,C_{\text{fail},\rho}}-1\,}. \]

  • Fréchet (\(\alpha>1\), scale \(s\); \(C_N=\sum_{k=1}^N\Gamma(k-\tfrac{1}{\alpha})/\Gamma(k)\)): with \(\mathrm{MV}_K\approx \dfrac{s\,C_N}{\alpha}K^{\frac{1}{\alpha}-1}\), \[ \boxed{\,K_\rho^\dagger =\left(\frac{s\,C_N}{\alpha\,\rho\,C_{\text{fail},\rho}}\right)^{\frac{\alpha}{\alpha-1}}.} \] Here \(s\) is the Fréchet scale parameter (units of impact per year), not a survival function. As \(\alpha\downarrow 1\) increases, \(K_\rho^\dagger\) explodes: in very heavy tails, very wide funnels can still be net positive.

With light tails, there’s a finite pool size after which turning up the hype (growing \(K\)) destroys net welfare. Every extra applicant burns \(C_{\text {fail},\rho}\) in private cost while adding \(\mathrm {MV}_K\) that shrinks like \(1/K\).

With heavy tails, the “wider funnel” argument can be correct—but only if the tail is actually heavy and the scale \(s\) is large.

For what is it worth, I think that it is possible that the distribution of impact levels of people is heavy-tailed, in some sense, in the real world. But I think it is less plausible that the distribution of applicants to AI safety roles is heavy-tailed. It seems to me that that 10x engineers and similar black swans are likely filtered out of that distribution by virtue of already having high-impact roles/and or sufficiently good connections to be head-hunted into high-impact roles rather than going via the sabbatical route.

Let’s plot total net welfare.

\[ \boxed{\,W(K)=B(K)\cdot H_\rho - \max\{K-N,0\}\cdot C_{\text{fail},\rho}\,} \] Where \(B(K)=\mathbb{E}[S_{N,K}]\) is in k$/year, \(H_\rho=1/\rho\) converts annual impact to present value, and \(C_{\text{fail},\rho}\) is in k$ (PV).

  • Exponential: exact \(B(K)=\dfrac{N}{\lambda}\Big(1+H_K-H_N\Big)\).
  • Fréchet (\(\alpha>1\)): large-\(K\) asymptotic \(B(K)\approx s\,C_N\,K^{1/\alpha}\) with \(C_N=\sum_{k=1}^N \dfrac{\Gamma\!\left(k-\tfrac{1}{\alpha}\right)}{\Gamma(k)}\).
  • We match the scales to Part 1 by setting the mean impact per hire per year to \(\mu_{\text{ref}}=22\): Exponential: \(1/\lambda=\mu_{\text{ref}}\Rightarrow \lambda=1/\mu_{\text{ref}}\). Fréchet: \(\mathbb{E}[I]=s,\Gamma(1-1/\alpha)=\mu_{\text{ref}}\Rightarrow s=\mu_{\text{ref}}/\Gamma(1-1/\alpha)\).
Code
# Field-level trade-off curves with discounting (PUBLIC MODEL)
# W(K) = B(K) * H_rho - max(K - N, 0) * C_fail_rho
# Where: H_rho = 1/(rho + mu), and C_fail_rho = c * (1 - exp(-rho*ell)) / rho

import numpy as np
import matplotlib.pyplot as plt
from scipy.special import gamma, digamma

# Optional styling: keep portable
try:
    from livingthing.matplotlib_style import set_livingthing_style
    set_livingthing_style()
except Exception:
    pass

# --- Parameters (aligned with Part 1) ---
N   = 24          # number of available roles
rho = 1/3.0       # discount rate per year  (~3-year effective horizon)
c   = 25.0        # k$/year private burn rate
ell = 0.5         # years of sabbatical runway
mu_ref = 22.0     # mean per-hire impact (k$/year, same Δu scale)
fail_cost_rho = c * (1 - np.exp(-rho * ell)) / rho  # PV of one failed pivot

# Impact distribution parameters
alphas = [2.0, 3.0, 4.0]  # Fréchet shapes (heavier→slower diminishing returns)
K_min, K_max = max(N, 20), 600
K = np.arange(K_min, K_max + 1)

EULER_GAMMA = 0.5772156649
def harmonic_number(n): return digamma(n + 1.0) + EULER_GAMMA

# --- Expected top-N impact under different tail assumptions ---
def B_exponential(K, N, mu_ref):
    """
    Impact per year (sum of top-N for Exp(λ) with mean 1/λ = mu_ref).
    Exact identity: B(K) = (N/λ) * (1 + H_K - H_N).
    """
    lam = 1.0 / mu_ref
    HK = harmonic_number(K.astype(float))
    HN = harmonic_number(float(N))
    return (N / lam) * (1.0 + HK - HN)

def C_N_frechet(N, alpha):
    j = np.arange(1, N + 1, dtype=float)
    return np.sum(gamma(j - 1.0 / alpha) / gamma(j))

def B_frechet_asymptotic(K, N, alpha, mu_ref):
    """
    Large-K asymptotic for sum of top-N of Fréchet(α, s), matched so E[I]=mu_ref.
    E[I] = s * Γ(1 - 1/α) => s = mu_ref / Γ(1 - 1/α).
    Then B(K) ≈ s * K^(1/α) * C_N.
    """
    s = mu_ref / gamma(1.0 - 1.0/alpha)
    Cn = C_N_frechet(N, alpha)
    return s * (K ** (1.0 / alpha)) * Cn

# --- Total discounted welfare: W(K) = PV(benefit) - PV(private burn) ---
def total_welfare(BK, K, N, fail_cost_rho, rho):
    H_rho = 1.0 / rho               # PV weight of one unit of annual impact
    failures = np.maximum(K - N, 0)
    return BK * H_rho - failures * fail_cost_rho

def argmax_idx(y):
    i = int(np.nanargmax(y))
    return i, float(y[i])

# --- Compute curves ---
B_exp = B_exponential(K, N, mu_ref)
W_exp = total_welfare(B_exp, K, N, fail_cost_rho, rho)
# Exponential: argmax
i_exp, Wexp_max = argmax_idx(W_exp)
Kexp_star = int(K[i_exp])

# Fréchet: argmax per alpha
frechet_results = []
for a in alphas:
    B_fr = B_frechet_asymptotic(K, N, a, mu_ref)
    W_fr = total_welfare(B_fr, K, N, fail_cost_rho, rho)
    i_star, Wmax = argmax_idx(W_fr)
    K_star = int(K[i_star])
    frechet_results.append((a, W_fr, K_star, Wmax))

Kexp_star = int(K[np.argmax(W_exp)])

# --- Plot ---
plt.figure(figsize=(8.5, 5.2))

# Exponential curve + argmax
plt.plot(K, W_exp, label="Exponential")
plt.scatter([Kexp_star], [Wexp_max], zorder=3)
plt.annotate(f"K*={Kexp_star}", (Kexp_star, Wexp_max),
             xytext=(6, 6), textcoords="offset points")

# Show where failures begin
plt.axvline(N, color="k", linestyle="--", lw=1)
plt.text(N, plt.ylim()[0], f"N={N}", rotation=90, va="bottom", ha="right")

# Fréchet curves + argmax
for (a, W_fr, K_star, Wmax) in frechet_results:
    plt.plot(K, W_fr, label=f"Fréchet α={a:.0f}")
    plt.scatter([K_star], [Wmax], zorder=3)
    plt.annotate(f"K*={K_star}", (K_star, Wmax),
                 xytext=(6, 6), textcoords="offset points")

plt.axhline(0, color="gray", linestyle="--", lw=1)
plt.xlabel("Applicant pool size K")
plt.ylabel("Total discounted welfare W(K) (k$ PV)")
plt.title("Total discounted welfare vs. pool size (single discount rate ρ)")
# Oversubscription thresholds (vertical guides)
lam = 1.0 / mu_ref                         # exponential mean = 1/lam = mu_ref
K_dagger_exp = (N / lam) * (1 / (rho * fail_cost_rho)) - 1

def K_dagger_frechet(alpha):
    s = mu_ref / gamma(1.0 - 1.0/alpha)    # Fréchet scale matched to mean
    Cn = C_N_frechet(N, alpha)
    base = (s * Cn) / (alpha * rho * fail_cost_rho)
    return base ** (alpha / (alpha - 1))

# Draw them (only if they’re in-range)
ymin, ymax = plt.ylim()
if K_min <= K_dagger_exp <= K_max:
    plt.axvline(K_dagger_exp, color="C0", ls=":", lw=1)
    plt.text(K_dagger_exp, ymin, "K† (Exp)", rotation=90, va="bottom", ha="center", color="C0")

for a in alphas:
    Kd = K_dagger_frechet(a)
    if K_min <= Kd <= K_max:
        plt.axvline(Kd, color="C1", ls=":", lw=1)
        plt.text(Kd, ymin, f"K† (α={a:.0f})", rotation=90, va="bottom", ha="center", color="C1")

plt.legend()
plt.grid(True, alpha=0.3)
plt.tight_layout()
plt.show()

# --- Print summary ---
print(f"ρ={rho:.2f}/yr (≈{1/rho:.1f}-yr horizon); fail_cost={fail_cost_rho:.2f} k$")
print(f"Exponential: K*={Kexp_star}, W(K*)={np.max(W_exp):.1f} k$")
for (a, W_fr, K_star, W_exp) in frechet_results:
    print(f"Fréchet α={a:.1f}: K*={K_star}, W(K*)={np.max(W_fr):.1f} k$")

ρ=0.33/yr (≈3.0-yr horizon); fail_cost=11.51 k$
Exponential: K*=137, W(K*)=3015.1 k$
Fréchet α=2.0: K*=248, W(K*)=3136.6 k$
Fréchet α=3.0: K*=73, W(K*)=1966.4 k$
Fréchet α=4.0: K*=43, W(K*)=1767.5 k$
  • This plots total net welfare \(W (K)\) (as opposed to marginal) and marks the argmax \(K^*\) for each family, i.e. the point where we’d stop widening the funnel if we cared about total welfare. The dashed line at \(K=N\) shows where failures begin: \((K>N\Rightarrow K-N\) people each pay \(C_{\text{fail},\rho})\). Note: the markers annotate \(K^*=\arg\max W(K)\); this generally differs from the oversubscription threshold \(K_\rho^\dagger\), which is defined by \(\mathrm{MV}_K H_\rho = C_{\text{fail},\rho}\).
  • Units: \(B(K)\) is in k$ per year and is converted to PV by multiplying by \(H_\rho=\frac{1}{\rho}\). The subtraction uses the discounted per-failure cost \(C_{\text{fail},\rho}=c\,\frac{1-e^{-\rho\ell}}{\rho}\).
  • Fréchet curves use the standard large-\(K\) asymptotic \(B(K)\approx s\,K^{1/\alpha}C_N\) (with \(s=\mu_{\text{ref}}/\Gamma(1-1/\alpha)\)). For exact finite-\(K\) values we could add a Monte Carlo estimator, but the large-\(K\) regime is what matters for oversubscription anyway.
  • We treat all future uncertainties about role duration, turnover, or project lifespan as already captured in the overall discount rate \(\rho\).

At the field level, Alice is one more data point on the dashboard; if the funnel is already redlining, her marginal impact may be negative.

Without evidence of heavy tails at your current applicant pool \(K\), widening the funnel likely increases social loss. Calibration beats enthusiasm.

3 Calibration playbook

The analysis suggests the AI safety field may be oversubscribed, causing a predictable loss for candidates (the Pivot Tax) and dissipating value for the field. If the goal is to optimize for welfare-maximizing matches rather than just maximizing applications, the ecosystem needs better feedback loops.

Think of the field as operating with two key gauges currently obscured:

  • Gauge A (Individual level): The candidate’s probability of success (\(p\)) relative to their break-even threshold (\(p_\star\)).
  • Gauge B (Field level): Whether the marginal value of new applicants (\(\mathrm {MV}_K\)) still exceeds their expected private costs (\(c\ell\)).

The responsibility for revealing these gauges rests primarily with the organizations that control the information flow: employers and advice organizations.

3.1 Improving field-level calibration (Gauge B)

To decide whether to keep widening the funnel, the field needs a clearer picture of how applicant numbers affect impact. The model highlights that this depends heavily on the distribution of candidate quality (the tail shape of \(F\)) and the private costs of pivoting (\(c\ell\)).

Estimating the tail shape of impact is notoriously difficult and there are no perfect measures. However, we are not even doing a good job at observing imperfect proxies and trends. Let’s consider some we could use:

  1. Estimate applicant costs (\(c\ell\)). Advice organizations or funders should survey applicants (both successful and unsuccessful) to estimate the typical time and financial costs incurred during a pivot attempt. This establishes the cost side of the equation.

  2. Track realized impact proxies. Employers should track proxies for the value-add of hires over time. Analyzing historical cohorts can help determine if widening the funnel is still yielding significantly better hires, or if returns are rapidly diminishing.

If returns are diminishing (as expected in light-tailed distributions) and applicant costs are high, the field should pause efforts to widen the generic funnel. Resources should instead shift towards creating more roles (\(N\)) or improving the impact of existing roles.

3.2 Improving individual-level calibration (Gauge A)

The most direct way to help individuals calibrate their decisions is to provide accurate base rates. This allows candidates to estimate their personal \(p\) and use the decision model presented in Part A.

  1. Publish stage-wise acceptance rates. Employers and fellowship programs should publish historical data on the number of applicants, interviews, and offers for different tracks and seniority levels. This is the single most impactful intervention for individual calibration.

  2. Provide informative feedback and rank. Base rates provide a population average, but candidates need personalized signals to update their estimate of \(p\). The highest-value information an employer can provide is an applicant’s approximate rank within the pool (e.g., “top quartile,” “middle 50%”) or standardized feedback based on initial screening.

    Providing this feedback is costly for employers — it requires more reviewer time and careful communication. However, these costs must be weighed against the significant deadweight loss incurred by miscalibrated candidates (the Pivot Tax). When organizations don’t pay the cost of providing feedback, they externalize that cost onto the applicants. Furthermore, if the field fails to provide this information, it risks a credibility cost by continuing to operate on opaque mechanisms. Investing in better feedback mechanisms internalizes the cost of calibration and reduces systemic deadweight loss.

Absent this data, advice organizations face a difficult challenge. They want to encourage talented people to enter the field, but without calibration, their encouragement may be misleading.

Some organizations already contribute valuable information. MATS (ML Alignment Theory Scholars), for example, publishes statistics about their application rounds. The 80,000 Hours job board provides context about the opportunity rate (\(r\)) and the scale of the field. Widespread adoption of this transparency is necessary for a calibrated ecosystem.

A potential fallback is for advisors to track and publish their own forecast calibration (e.g., Brier scores) regarding candidate success. However, this is a second-best solution. It provides a noisy signal and places the burden of interpretation on the candidates, who may not be resourced to evaluate such scores. It does not replace the need for ground-truth base rates from employers. If an advice organization does not track outcomes, its advice cannot be calibrated except by coincidence.

Publishing base rates and estimating applicant costs converts guesswork into readable dials. Providing individualized feedback sharpens those readings significantly. When the gauges suggest oversubscription, the priority should shift from widening the funnel to improving selection and signaling.

NoteHow close the model maps to reality

This is a deliberately simple model. Applications are rarely independent, timelines stretch or shorten, and organisations cap capacity long before theory predicts. Treat the math as a calibration aid, not as ground truth. When in doubt, default to the conservative reading: the system is probably less forgiving than the equations suggest.

4 Normative implications

The lack of calibration mechanisms in the AI safety hiring pipeline has distinct implications for different actors in the ecosystem.

For mid-career individuals, the decision to pivot is high-stakes. The model in Part A provides a framework for making this decision under uncertainty. The key takeaway is that a personal break-even probability (\(p_\star\)) is sensitive to opportunity costs (\(c\)) and expected impact gain (\(\Delta u\)). Without strong, personalized evidence suggesting a probability of success (\(p\)) is above this threshold, a pivot attempt involving significant unpaid time is likely EV-negative. Candidates should use the model to calculate their own \(p_\star\) and seek cheap signals of fit—such as applying to a few roles before leaving their current job—before committing significant resources.

For early-career individuals, the opportunity costs (\(c\)) are generally lower and the time horizon (\(T\)) longer, making the gamble more favorable. However, the fundamental problem of estimating \(p\) remains.

For advice organizations and employers, the implications are systemic. The current system incentivizes maximizing visible proxies, such as application volume, while the true goal—welfare-maximizing matches at low opportunity cost—is under-observed. This is a classic setup for Goodhart’s Law, predictably leading to over-entry and the dissipation of value.

The credibility of advice depends on its calibration to reality. The system is misaligned because the organizations influencing the funnel size do not internalize the costs borne by unsuccessful applicants. Until organizations provide evidence-based estimates of success probabilities, their generic encouragement to “do AI safety stuff” should be treated with scepticism by mid-career professionals.

A note on impact valuation (\(i_1\)): The model assumes a certain impact gain from the pivot. However, in a saturated field, the marginal impact of an individual depends on whether they are significantly better than the next-best candidate who would have taken the role otherwise. This counterfactual impact is hard to estimate, which adds another layer of uncertainty and emphasises the need for calibration on more observable metrics like success probability and cost.

5 Asks and next steps

The Pivot Tax is a deadweight loss resulting primarily from poor information flow. We can reduce this loss by implementing concrete mechanisms for transparency and calibration. The primary focus must be on improving the information environment.

5.1 Improving the information environment

The following actions would significantly improve the ability of both individuals and the field to calibrate decisions.

  1. Stage-count reports. Employers should commit to publishing quarterly or annual summaries of application funnels, detailing the number of applicants, candidates reaching each interview stage, and hires, segmented by track and seniority. A simple standardized format (e.g., CSV) would suffice. This is the single most important step toward anchoring realistic success probabilities (\(p\)).

  2. Standardized feedback and ranking. Employers should develop mechanisms to provide standardized feedback or an indication of relative rank to applicants, even those rejected early. While resource-intensive, this provides a crucial personalized signal of fit. The field must recognize that the cost of providing this information should be weighed against the collective cost of the Pivot Tax and the long-term credibility of the field.

  3. Applicant cost surveys. Advice organizations should regularly survey the community to estimate the typical private costs (\(c\ell\)) associated with pivot attempts, including foregone wages and time spent. Publishing these estimates helps determine the field-level oversubscription threshold (\(K^\dagger\)).

We can try to get the information ourselves by gossip and guesstimation, but this is a poor substitute for systematic transparency.

5.2 Secondary improvements: Reducing private costs

Beyond calibration, there are ways to reduce the deadweight loss by lowering the private costs of pivoting or by better utilizing existing signals.

  • Utilizing transition grants as signals. Organizations like Open Philanthropy offer career transition grants. These play a dual role in the ecosystem, serving both as funding and as information gates.
    • If received, a grant directly lowers the private cost of pivoting (\(c\)) for the candidate, making the gamble more favorable by reducing the financial barrier to exploration.
    • If denied, it serves as a valuable, albeit noisy, calibration signal. The grantmaker’s assessment, presumably calibrated by observing many such transitions, offers an external signal about the candidate’s prospects. If a major funder declines to underwrite the transition, candidates should use this information to significantly update their estimate of \(p\) downwards, recognizing that their sabbatical will not only be more costly but also less likely to succeed.
  • Mechanism design (Soft caps). In capacity-constrained hiring rounds or fellowships, implementing soft caps—where the application window automatically pauses after a certain number of applications—can reduce excessive congestion. Experimental evidence suggests this can reduce applicant-side waste without significantly harming match quality (J. J. Horton et al. 2024).

Transparency on base rates, individualized feedback, and applicant costs is essential for a healthy ecosystem. Candidates should also use existing mechanisms, such as transition grants, as calibration tools, not just funding sources. Once this information is available and correctly interpreted, candidates can make informed decisions, and the predictable losses associated with miscalibrated pivots will shrink.

All of which is to say, if you don’t get that Open Phil transition grant, don’t quit your current job.

6 Coda: The Donation Baseline

Throughout this analysis, we have modeled the career pivot as a gamble with associated private costs (\(c\ell\)) and potential impact gains (\(\Delta u\)). The central argument is that without proper calibration, this gamble often results in a deadweight loss—the Pivot Tax.

For candidates motivated primarily by impact, we need to compare the pivot strategy against a readily available alternative: remaining in the current role and donating the resources that would have been consumed by the pivot attempt.

The private cost of the sabbatical, \(c\ell\), represents time and money invested in the transition. In Alice’s worked example, this cost was estimated at $12.5k over six months. If her pivot fails, that value is dissipated. However, she could instead choose to donate that $12.5k directly to AI safety organizations.

This reframes the decision. The comparison is not simply between the expected value of the pivot and the status quo. It is between the expected value of the pivot and the certain impact of donating the pivot costs.

If the EV calculation for the pivot is marginal, negative, or highly uncertain due to poor information (an opaque \(p\)), the donation alternative—a form of “Earning to Give”—provides a guaranteed impact without the associated career risk. Even if the EV is slightly positive, the certainty of the donation might be preferable when accounting for risk aversion (which, for simplicity, was excluded from the main model).

This baseline reinforces the systemic importance of calibration. When individuals undertake EV-negative pivots based on uncalibrated advice, the loss is not merely a private cost to the individual. It represents a destruction of potential impact that could have been reliably realized through donations. A calibrated ecosystem ensures that resources—whether time, talent, or capital—are allocated where they generate the most value, rather than dissipated in poorly informed gambles.

7 Where next?

I really need to calculate the field-wise deadweight loss from this misalignment. How many people have produced a net negative impact on society by burning \(c\) instead of donating \(d\) because they miscalibrated pivots for negligible change in \(I\)? However, I’ve already burned more time than I could spare on this, so consider that tabled for later.

I fed this essay to an LLM for feedback. It suggested I discuss congestion costs for employers. After due consideration, I disagree. There might be second-order congestion costs, but I don’t think first-order effects are significant. Generally, employers who have filled a given role can just ignore excess applications, and there’s a lot of evidence to suggest that they do so (J. Horton, Kerr, and Stanton 2017; J. Horton and Vasserman 2021; J. J. Horton et al. 2024). But maybe some employers can tell me if I’m wrong.

More generally, I’d like feedback from people deeper in the AI safety career ecosystem. I’d love to chat with people from 80,000 Hours, MATS, FHI, CHAI, Redwood Research, Anthropic, etc., about this. What have I got wrong? What have I missed? I’m open to the possibility that this is well understood and being actively managed behind the scenes, but I haven’t seen it laid out this way anywhere.

8 Further reading

Resources that complement the mechanism-design view of the AI safety career ecosystem:

9 References

Arulampalam. 2000. Is Unemployment Really Scarring? Effects of Unemployment Experiences on Wages.”
Caron, Teh, and Murphy. 2014. Bayesian Nonparametric Plackett–Luce Models for the Analysis of Preferences for College Degree Programmes.” The Annals of Applied Statistics.
Earnest, Allen, and Landis. 2011. Mechanisms Linking Realistic Job Previews with Turnover: A Meta‐Analytic Path Analysis.” Personnel Psychology.
Hill, Yin, Stein, et al. 2025. The Pivot Penalty in Research.” Nature.
Horton, John, Kerr, and Stanton. 2017. Digital Labor Markets and Global Talent Flows.” Working Paper. Working Paper Series.
Horton, John J, Sloan, Vasserman, et al. 2024. Reducing Congestion in Labor Markets: A Case Study in Simple Market Design.”
Horton, John, and Vasserman. 2021. Job-Seekers Send Too Many Applications: Experimental Evidence and a Partial Solution.” 021.
Schmidt, Frank L., and Hunter. 1998. The Validity and Utility of Selection Methods in Personnel Psychology: Practical and Theoretical Implications of 85 Years of Research Findings. Psychological Bulletin.
Schmidt, F. L., Oh, and Shaffer. 2016. The Validity and Utility of Selection Methods in Personnel Psychology: Practical and Theoretical Implications of 100 Years of Research.”
Skaperdas. 1996. Contest Success Functions.” Economic Theory.

Footnotes

  1. To be consistent we’ll need to take this to be a local linear approximation at your current wage and impact level, but that’s fine for a simple decision.↩︎