Denkenberger
Model
A few of us at ALLFED (myself, @jamesmulhall, and others) have been thinking about response planning for essential (vital) workers in extreme pandemics. Our impression is that there's a reasonable chance we will not be prepared for an extreme pandemic if it happens, so we should have back-up plans in place to keep basic services functioning and prevent collapse. We think this is probably a neglected area that more people should be working on, and we're interested in whether others think this is likely to be a high-impact topic. We decided to compare it to a standard and evidence-backed intervention to protect the vital workforce that is receiving funding from EA — stockpiling of pandemic proof PPE (P4E).
We asked Squiggle AI to create two cost-effectiveness analyses comparing stockpiling P4E vs research and planning to rapidly scale up after the outbreak transmission-reducing interventions (e.g. UV) to keep essential workers safe. Given the additional costs of both interventions could be significantly lowered by influencing funding governments have already allocated to stockpiling/response planning, we ran the model with (linked here) and without a message (linked here) to only consider the costs of philanthropic funding.
Summary result:
Considering all spending, research and planning is estimated as 34 (8.5–140) times as cost-effective as stockpiling
Considering only philanthropic spending, research and planning is estimated as 47 (23–100) times as cost-effective as stockpiling
We did not feed any numbers into the model, but the ones it self generated seemed reasonably sensible (e.g., Kevin Esvelt's quote of $20 billion for stockpiling adequate PPE for the US falls within the $4-20 billion estimate by the model)
Prompt:
Create a cost-effectiveness analysis comparing two interventions to keep US essential workers safe in a pandemic with extremely high transmissibility and fatality rates. Assess the interventions on the probability they are successful at preventing the collapse of civilization. Only include money spent before the pandemic happens as there will be plenty of money available for implementation after it starts.
1: Stockpiling elastomeric half mask respirators and PAPRs before the extreme pandemic.
2: Researching and planning to scale up transmission reduction interventions rapidly after the pandemic starts, including workplace adaptations, indoor air quality interventions (germicidal UV, in-room filtration, ventilation), isolation of workers in on-site housing, and contingency measures for providing basic needs if infrastructure fails.
Outputs:
-
narrative and explanations of the logic behind all of the numbers used
-
ranges of costs for the two options
-
ranges of effectiveness for the two options
-
cost-effectiveness for the two options
-
mean and median ratios of cost effectiveness of planning vs stockpiling
-
distribution plots of the cost effectiveness of planning vs stockpiling
Optional message:
Important: only account for philanthropic funding costs to make these interventions happen. Assume that governments already have pandemic preparedness funding allocated for stockpiles and response planning. This may reduce philanthropic costs if stockpiling interventions can redirect government purchases from disposable masks to more protective elastomeric respirators/PAPRs or if research and planning interventions can add their recommendations to existing government frameworks to prepare essential industries for disasters.
Attachments:
[1] https://squigglehub.org/models/ALLFED/pandemics-comparison-philanthropic-costs
/*
Generated by Squiggle AI. Workflow ID: d1b7c4c0-8c31-411b-8c34-5ab95b85b610
*/
import "hub:ozziegooen/sTest" as sTest
@name("Input Parameters")
inputs = {
@name("US Essential Workers")
@doc("According to BLS, ~50M essential workers in US")
@doc(
"Base estimate 45-55M from BLS, with 10% chance of 40-60M range due to:
- Definition variations of 'essential' workers
- Informal/gig economy workers
- Seasonal variations"
)
essentialWorkers = mx([45M to 55M, 40M to 60M], [0.9, 0.1])
@name("Extreme Pandemic Parameters")
pandemicParams = {
@name("Infection Fatality Rate")
@doc("Much higher than COVID-19 due to extreme scenario")
@doc(
"Base estimate 5-15% for highly lethal pandemic, with 10% chance of 15-30% for catastrophic scenarios involving:
- Multiple strains
- Healthcare system collapse
- Supply chain disruptions affecting medical care"
)
ifr = mx([5% to 15%, 15% to 30%], [0.9, 0.1])
@name("R0 Value")
@doc("Basic reproduction number, highly transmissible")
@doc(
"Base R0 of 4-8 (higher than COVID), with 15% chance of 8-15 due to:
- Potential aerosol transmission
- Indoor workplace exposure
- Long infectious period"
)
r0 = mx([4 to 8, 8 to 15], [0.85, 0.15])
@name("Probability of Extreme Pandemic in 10 Years")
@doc("Low probability but catastrophic impact event")
@doc(
"Base probability 0.5-2% over 10 years, with 5% chance of 2-5% due to:
- Increasing lab work on dangerous pathogens
- Climate change effects on disease emergence
- Increasing global travel and trade"
)
probability = mx([0.5% to 2%, 2% to 5%], [0.95, 0.05])
{ ifr, r0, probability }
}
@name("Stockpiling Intervention")
stockpiling = {
@name("Cost per Worker Protected")
@doc(
"Includes elastomeric respirators (~$40) + PAPRs for high-risk (~$800)"
)
@format("$,.0f")
@doc(
"Costs highly variable depending on:
- Bulk purchase discounts
- Mix of respirator types needed
- Storage and maintenance costs
- Training and fit-testing costs"
)
costPerWorker = mx([60 to 120, 40 to 200], [0.9, 0.1])
@name("Coverage Rate")
@doc("Percentage of essential workers who can be covered")
@format(".0%")
@doc(
"Coverage varies significantly based on:
- Worker acceptance and compliance
- Fit-testing success rates
- Supply chain capacity
- Storage constraints"
)
coverage = mx([40% to 70%, 20% to 90%], [0.8, 0.2])
@name("Effectiveness at Preventing Death")
@doc("When properly used, these devices are highly effective")
effectiveness = 85% to 95%
@name("Implementation Success Rate")
@doc("Challenges with training, fit-testing, and compliance")
implementationSuccess = 60% to 80%
{ costPerWorker, coverage, effectiveness, implementationSuccess }
}
@name("Planning & Research Intervention")
planning = {
@name("Total Research & Planning Costs")
@doc("Studies, guidelines, protocols, training materials")
totalCost = 40M to 120M
@name("Coverage Rate")
@doc("Percentage of essential workers reached")
coverage = 60% to 90%
@name("Effectiveness at Preventing Death")
@doc("Combined effect of multiple interventions")
effectiveness = 70% to 90%
@name("Implementation Success Rate")
@doc("Easier to implement than stockpiling")
implementationSuccess = 75% to 90%
{ totalCost, coverage, effectiveness, implementationSuccess }
}
{ essentialWorkers, pandemicParams, stockpiling, planning }
}
@name("Stockpiling Analysis")
stockpilingAnalysis = {
totalCost = inputs.essentialWorkers * inputs.stockpiling.costPerWorker *
inputs.stockpiling.coverage
workersSaved = inputs.essentialWorkers * inputs.stockpiling.coverage *
inputs.stockpiling.effectiveness *
inputs.stockpiling.implementationSuccess *
inputs.pandemicParams.ifr *
inputs.pandemicParams.probability
costPerLifeSaved = totalCost / workersSaved
{ totalCost, workersSaved, costPerLifeSaved }
}
@name("Planning Analysis")
planningAnalysis = {
totalCost = inputs.planning.totalCost
workersSaved = inputs.essentialWorkers * inputs.planning.coverage *
inputs.planning.effectiveness *
inputs.planning.implementationSuccess *
inputs.pandemicParams.ifr *
inputs.pandemicParams.probability
costPerLifeSaved = totalCost / workersSaved
{ totalCost, workersSaved, costPerLifeSaved }
}
@name("Comparative Analysis")
comparison = {
@name("Cost-Effectiveness Ratio")
@doc("Ratio of planning cost-effectiveness to stockpiling cost-effectiveness")
ratio = stockpilingAnalysis.costPerLifeSaved /
planningAnalysis.costPerLifeSaved
@name("Summary Statistics")
stats = { meanRatio: mean(ratio), medianRatio: quantile(ratio, 0.5) }
{ ratio, stats }
}
// Move import to top
analysis_tests = sTest.describe(
"Analysis Tests",
[
sTest.test(
"stockpiling total cost is reasonable",
{|| sTest.expect(mean(stockpilingAnalysis.totalCost)).toBeLessThan(10B)}
),
sTest.test(
"planning is generally more cost-effective",
{|| sTest.expect(mean(comparison.ratio)).toBeGreaterThan(1)}
),
sTest.test(
"saved workers are less than total workers",
{
||
sTest.expect(mean(planningAnalysis.workersSaved)).toBeLessThan(
mean(inputs.essentialWorkers)
)
}
),
]
)
@notebook
@startOpen
summary = [
"## Pandemic Intervention Analysis Summary
This model compares two interventions for protecting essential workers during an extreme pandemic:
1. Pre-pandemic stockpiling of advanced respiratory protection
2. Pre-pandemic research and planning for rapid implementation
### Key Findings",
"#### Stockpiling Intervention
- Total Cost: **$" +
String(mean(stockpilingAnalysis.totalCost) / 1B, ".1f") +
"B** (range: $" +
String(quantile(stockpilingAnalysis.totalCost, 0.05) / 1B, ".1f") +
"B to $" +
String(quantile(stockpilingAnalysis.totalCost, 0.95) / 1B, ".1f") +
"B)",
"- Potential Lives Saved: **" +
String(mean(stockpilingAnalysis.workersSaved) / 1000, ".0f") +
"k**",
"- Cost per Life Saved: **$" +
String(mean(stockpilingAnalysis.costPerLifeSaved) / 1000, ".0f") +
"k**",
"#### Planning Intervention
- Total Cost: **$" +
String(mean(planningAnalysis.totalCost) / 1M, ".0f") +
"M**",
"- Potential Lives Saved: **" +
String(mean(planningAnalysis.workersSaved) / 1000, ".0f") +
"k**",
"- Cost per Life Saved: **$" +
String(mean(planningAnalysis.costPerLifeSaved) / 1000, ".0f") +
"k**",
"#### Comparative Analysis",
"- Mean Cost-Effectiveness Ratio: **" +
String(comparison.stats.meanRatio, ".1f") +
"x**",
"- Median Cost-Effectiveness Ratio: **" +
String(comparison.stats.medianRatio, ".1f") +
"x**",
"### Cost-Effectiveness Comparison",
comparison.ratio,
"### Major Uncertainties
- **Pandemic Probability**: **" +
String(Dist.inv(inputs.pandemicParams.probability, 0.05) * 100, ".1f") +
"%** to **" +
String(Dist.inv(inputs.pandemicParams.probability, 0.95) * 100, ".1f") +
"%** chance over 10 years
- **Implementation Success**: **" +
String(
Dist.inv(inputs.stockpiling.implementationSuccess, 0.05) * 100,
".1f"
) +
"%** to **" +
String(
Dist.inv(inputs.stockpiling.implementationSuccess, 0.95) * 100,
".1f"
) +
"%** for stockpiling
- **Coverage Achievable**: Stockpiling reaches **" +
String(Dist.inv(inputs.stockpiling.coverage, 0.05) * 100, ".1f") +
"%** to **" +
String(Dist.inv(inputs.stockpiling.coverage, 0.95) * 100, ".1f") +
"%** of workers
- **Cost Uncertainty**: Stockpiling costs **$" +
String(Dist.inv(inputs.stockpiling.costPerWorker, 0.05), ",.0f") +
"** to **$" +
String(Dist.inv(inputs.stockpiling.costPerWorker, 0.95), ",.0f") +
"** per worker
### Key Takeaways
1. Planning intervention appears more cost-effective due to lower upfront costs and broader coverage
2. Stockpiling provides more certain protection but at higher cost
3. Both interventions show positive expected value given catastrophic potential
",
]
[2] https://squigglehub.org/models/ALLFED/pandemics-comparison-full-costs
/*
Generated by Squiggle AI. Workflow ID: 18590fe1-76ef-4b85-b252-a551010ad3aa
*/
import "hub:ozziegooen/sTest" as sTest
// == Model Parameters ==
@doc("Number of essential workers in the US needed to prevent collapse")
essentialWorkers = 10M to 30M
@doc("Cost per respirator+filter stockpile for 2 years ($)")
respiratorCost = 100 to 300
@doc("Number of respirators needed per worker (includes spares/replacements)")
respiratorsPerWorker = 2 to 4
@doc("Cost to research & plan interventions ($)")
planningCost = {
research = 20M to 100M
suppliesAssessment = 10M to 30M
facilityPlanning = 20M to 80M
contingencyPlanning = 30M to 120M
research + suppliesAssessment + facilityPlanning + contingencyPlanning
}
// == Intervention Costs ==
@name("Total Stockpiling Cost ($)")
@format("$,.0f")
stockpilingCost = essentialWorkers * respiratorCost * respiratorsPerWorker
// == Effectiveness Parameters ==
@doc("Probability of civilization-threatening pandemic in next 20 years")
pandemicProb = 0.01 to 0.1
@doc("Probability stockpiling prevents collapse given pandemic")
stockpilingEffectiveness = {
protectionLevel = mx([0.7 to 0.95, 0.4 to 0.99], [0.9, 0.1]) // Protection level with uncertainty
complianceRate = mx([0.5 to 0.8, 0.2 to 0.9], [0.9, 0.1]) // Compliance with high uncertainty
logisticsSuccess = mx([0.6 to 0.9, 0.1 to 0.95], [0.9, 0.1]) // Logistics with crisis uncertainty
protectionLevel * complianceRate * logisticsSuccess
}
@doc("Probability planning prevents collapse given pandemic")
planningEffectiveness = {
interventionEfficacy = mx([0.4 to 0.8, 0.1 to 0.9], [0.9, 0.1]) // Combined measures with uncertainty
implementationSuccess = mx([0.3 to 0.7, 0.05 to 0.85], [0.9, 0.1]) // Implementation with crisis uncertainty
interventionEfficacy * implementationSuccess
}
// == Cost Effectiveness Analysis ==
@doc("Cost effectiveness of stockpiling (lives saved per $)")
stockpilingCE = pandemicProb * stockpilingEffectiveness / stockpilingCost
@doc("Cost effectiveness of planning (lives saved per $)")
planningCE = pandemicProb * planningEffectiveness / planningCost
@doc("Ratio of planning CE to stockpiling CE")
planningToStockpilingRatio = planningCE / stockpilingCE
// == Tests ==
model_tests = sTest.describe(
"Model Tests",
[
sTest.test(
"stockpiling cost is reasonable",
{|| sTest.expect(mean(stockpilingCost)).toBeLessThan(1T)}
),
sTest.test(
"effectiveness probabilities are valid",
{|| sTest.expect(mean(stockpilingEffectiveness)).toBeLessThan(1)}
),
sTest.test(
"CE ratio is calculated",
{|| sTest.expect(mean(planningToStockpilingRatio)).toBeGreaterThan(0)}
),
]
)
// == Summary Output ==
@notebook
@startOpen
summary = [
"## Pandemic Intervention Cost-Effectiveness Analysis",
"### Cost Ranges",
"- Stockpiling: **$" + String(Dist.quantile(stockpilingCost, 0.05), ",.0f") +
"** to **$" +
String(Dist.quantile(stockpilingCost, 0.95), ",.0f") +
"**",
"- Planning: **$" + String(Dist.quantile(planningCost, 0.05), ",.0f") +
"** to **$" +
String(Dist.quantile(planningCost, 0.95), ",.0f") +
"**",
"### Effectiveness Given Pandemic",
"- Stockpiling: **" + String(mean(stockpilingEffectiveness) * 100, ".1f") +
"%** chance of preventing collapse",
"- Planning: **" + String(mean(planningEffectiveness) * 100, ".1f") +
"%** chance of preventing collapse",
"### Cost-Effectiveness Comparison",
"Planning vs Stockpiling Ratio:",
"- Mean: **" + String(mean(planningToStockpilingRatio), ".2f") + "x**",
"- Median: **" + String(quantile(planningToStockpilingRatio, 0.5), ".2f") +
"x**",
"### Key Uncertainties",
"- Essential worker population needed",
"- Pandemic probability and severity",
"- Implementation success under crisis conditions",
"- Actual protective capability of interventions",
"### Visualizations",
Plot.dists([stockpilingCE, planningCE]),
planningToStockpilingRatio,
]
Evaluation Results
Run 1:
Based on the provided Fermi model submission, I would rate the surprisingness of the key findings as a 4 out of 10. Here's my reasoning:
-
Contradiction of Expectations: The finding that research and planning is more cost-effective than stockpiling PPE is not entirely surprising. Many in the rationalist and EA communities already emphasize the importance of preparedness and planning over purely material solutions.
-
Counterintuitiveness: The magnitude of the difference (34-47 times more cost-effective) is somewhat counterintuitive and might be considered mildly surprising. However, large differences in cost-effectiveness between interventions are not uncommon in EA analyses.
-
Discovery of Unknowns: The model doesn't seem to uncover any previously unrecognized issues or risks. It mainly compares two known approaches to pandemic preparedness.
-
Magnitude of Difference: The 34-47 times difference in cost-effectiveness is significant, but not unprecedented in EA cost-effectiveness comparisons.
Specific details that illustrate the mildly surprising aspects:
-
The model suggests that even when considering full costs (not just philanthropic spending), research and planning is still substantially more cost-effective than stockpiling. This challenges the intuition that material preparations (like PPE) would be more directly impactful.
-
The robustness of the result across different cost scenarios (full costs vs. philanthropic costs) adds some weight to the surprising nature of the finding.
-
The model's self-generated numbers aligning with external estimates (like Kevin Esvelt's quote) lends credibility to the results, which might make the large difference in cost-effectiveness more surprising.
However, several factors reduce the overall surprisingness:
-
The EA community often finds large differences in cost-effectiveness between interventions, so the magnitude itself is not highly surprising.
-
The importance of research and planning is already well-recognized in EA circles, especially for low-probability, high-impact scenarios like extreme pandemics.
-
The model doesn't challenge any core assumptions or beliefs held by the rationalist or EA communities.
In conclusion, while the specific numbers and magnitude of difference might raise some eyebrows, the overall direction and implications of the findings align relatively well with existing EA thinking on pandemic preparedness. Therefore, a rating of 4 out of 10 on the surprisingness scale seems appropriate.
Run 2:
Based on the provided Fermi model submission, I would rate the surprising nature of the findings as a 6 out of 10. Here's my reasoning:
-
Contradiction of Expectations: The finding that research and planning is significantly more cost-effective than stockpiling PPE is somewhat surprising. Many people might intuitively expect that having physical resources on hand (like PPE) would be more effective than planning alone. This contradicts the common "preparedness" mindset.
-
Counterintuitiveness: The magnitude of the difference in cost-effectiveness is counterintuitive. The model suggests that planning is 34-47 times more cost-effective than stockpiling, which is a substantial difference that might not be immediately apparent to most people.
-
Discovery of Unknowns: The model doesn't necessarily uncover completely unknown issues, but it does highlight the potential undervaluation of research and planning in pandemic preparedness, which might not be widely recognized.
-
Magnitude of Difference: The large difference in cost-effectiveness (34-47 times) is significant and would likely be surprising to many in the rationalist and EA communities.
Specific details that illustrate the surprising aspects:
-
The model suggests that even when considering all spending (not just philanthropic), planning is still 34 times more cost-effective than stockpiling. This challenges the notion that government stockpiling is the most efficient way to prepare for pandemics.
-
The finding that philanthropic spending on planning could be 47 times more cost-effective than stockpiling suggests a potentially high-impact opportunity that may not be widely recognized in the EA community.
-
The model's self-generated numbers aligning with expert estimates (like Kevin Esvelt's $20 billion quote) lends credibility to the surprising results.
While these findings are certainly noteworthy, they don't completely upend our understanding of pandemic preparedness. The general idea that planning and research are important is not new, even if the magnitude of their importance is surprising. Additionally, the EA and rationalist communities are often open to counterintuitive findings based on careful analysis, which slightly reduces the surprise factor.
Therefore, a score of 6 out of 10 seems appropriate, indicating that the findings are more surprising than average, but not completely revolutionary or paradigm-shifting.
Run 3:
To evaluate the level of surprise for the rationalist and effective altruism (EA) communities regarding this Fermi model, I'll consider the key findings and rate them on the 0-10 scale:
- Cost-effectiveness comparison: The model suggests that research and planning is 34-47 times more cost-effective than stockpiling PPE. This is quite surprising, as many in the EA community might have assumed stockpiling to be more directly impactful. The magnitude of difference is significant.
Rating: 7/10
- Philanthropic vs. full costs: The finding that considering only philanthropic spending makes planning even more cost-effective (47x vs 34x) is somewhat counterintuitive. This suggests that leveraging existing government funding can amplify the impact of philanthropic efforts.
Rating: 6/10
- Scale of impact: The model deals with preventing the collapse of civilization, which is a key concern in EA circles. However, the fact that relatively modest investments in planning could have such a large impact on this existential risk might be surprising.
Rating: 5/10
- Uncertainty and ranges: The wide ranges in the results (e.g., 8.5-140 times more cost-effective) reflect significant uncertainty, which is expected in such models but might be surprising in its extent.
Rating: 3/10
- Neglectedness: The assertion that this is a neglected area aligns with EA thinking about cause prioritization, but the degree of neglect compared to the potential impact might be surprising.
Rating: 4/10
Overall rating: 6/10
This rating reflects that while some aspects of the model align with EA and rationalist thinking, the magnitude of the difference in cost-effectiveness and the potential impact of planning versus stockpiling is likely to be quite surprising to many in these communities. The model challenges some common assumptions about pandemic preparedness and highlights a potentially overlooked approach to addressing extreme risks.
The high cost-effectiveness of planning compared to stockpiling, especially when considering philanthropic spending, is the most surprising element. This finding could potentially shift priorities in pandemic preparedness efforts within these communities.
Run 1:
Based on the provided Fermi model submission, I would rate the importance of this subject matter to the rationalist and effective altruism communities as an 8 out of 10.
Rationale:
-
Relevance (9/10):
- The model directly addresses pandemic preparedness, which is a central concern in existential risk reduction - a key focus area for both rationalists and effective altruists.
- It compares two specific interventions (stockpiling vs. planning/research) that are highly relevant to pandemic mitigation strategies.
- The model considers the impact on essential workers and the potential for civilizational collapse, which aligns with long-termist perspectives common in these communities.
-
Impact Potential (7/10):
- The findings could significantly influence decision-making and priority-setting within these communities, especially regarding resource allocation for pandemic preparedness.
- The cost-effectiveness analysis provides actionable insights that could guide funding decisions and policy recommendations.
- The model's results suggest that planning and research interventions may be substantially more cost-effective than stockpiling, which could shift focus and resources within the EA community.
-
Additional Considerations:
- The model uses Squiggle AI, a tool developed within the EA community, demonstrating integration with existing EA methodologies.
- The analysis considers both philanthropic and full-cost scenarios, which is relevant for both individual EA donors and policy-makers.
- The topic of extreme pandemics and civilizational resilience is highly aligned with existential risk concerns prevalent in these communities.
Overall, this model addresses a crucial topic for the rationalist and EA communities, providing valuable insights that could significantly impact decision-making and resource allocation. The high relevance of the subject matter and the potential for substantial impact on community priorities justify the 8/10 rating.
Run 2:
Based on the provided Fermi model submission, I would rate the importance of this model's subject matter to the rationalist and effective altruism communities as an 8 out of 10.
Rationale:
-
Relevance (9/10):
- The model directly addresses a critical issue for both rationalist and EA communities: pandemic preparedness and prevention of civilizational collapse.
- It aligns well with EA priorities of existential risk reduction and global catastrophic risk mitigation.
- The focus on essential workers and maintaining basic services during an extreme pandemic is highly relevant to preserving societal function.
-
Impact Potential (7/10):
- The model's findings could significantly influence decision-making and priority-setting within these communities, especially regarding resource allocation for pandemic preparedness.
- The comparison between stockpiling and research/planning interventions provides actionable insights for policymakers and EA funders.
- The cost-effectiveness analysis offers a quantitative basis for comparing interventions, which is valued in both rationalist and EA circles.
Overall Score: 8/10
This score reflects the high relevance of the topic and the potential for meaningful impact on decision-making within these communities. The model addresses a neglected area of pandemic preparedness, which is particularly valued in EA circles. The quantitative approach and cost-effectiveness comparison align well with rationalist principles of evidence-based decision-making.
While not a perfect 10, the score of 8 acknowledges that while this is a highly important topic, there may be other competing priorities within these communities (such as AI safety or other existential risks) that could be considered equally or more pressing by some members.
Run 3:
Based on the provided Fermi model submission, I would rate the importance of this model's subject matter to the rationalist and effective altruism communities as an 8 out of 10.
Justification:
-
Relevance (9/10):
- The model directly addresses pandemic preparedness, which is a core concern within existential risk reduction - a key focus area for both rationalists and effective altruists.
- It compares two specific interventions (stockpiling vs. planning/research) that are highly relevant to pandemic mitigation strategies.
- The model considers the potential for civilization collapse, which aligns with long-termist perspectives common in these communities.
-
Impact Potential (7/10):
- The findings could significantly influence decision-making and resource allocation within these communities, particularly for organizations focused on global catastrophic risks.
- The cost-effectiveness comparison between stockpiling and planning/research provides actionable insights that could shape priorities in pandemic preparedness efforts.
- The model's consideration of philanthropic funding vs. government spending aligns well with how many EA organizations approach intervention planning.
Additional considerations:
- The model uses Squiggle AI, a tool developed within the EA community, demonstrating its integration with existing EA methodologies.
- The focus on essential workers and maintaining basic services during an extreme pandemic scenario is a novel angle that could spark further research and discussion within these communities.
- The model's explicit consideration of uncertainty and use of probabilistic reasoning aligns well with rationalist approaches to decision-making.
While not scoring a perfect 10, this model addresses a highly relevant topic with significant potential impact on decision-making within the rationalist and EA communities. Its approach and focus areas are well-aligned with the interests and methodologies of these groups, making it an important contribution to the field of pandemic preparedness and existential risk reduction.
Run 1:
Thank you for providing this detailed Fermi model submission on pandemic interventions. I'll evaluate the robustness of the model's key findings based on the criteria you've outlined.
Evaluation:
- Sensitivity to Assumptions: The model incorporates uncertainty in many key parameters, which is a strength. For example, it uses mixture distributions for pandemic probability, infection fatality rate, and R0 value. This approach captures both baseline estimates and more extreme scenarios.
However, some assumptions could significantly impact results if changed:
- The definition and number of "essential workers"
- The effectiveness of interventions in preventing deaths
- Implementation success rates
The model would be more robust if it included sensitivity analyses for these key parameters.
- Evidence Base: The model draws on some credible data sources, such as BLS estimates for essential workers. However, many parameters are based on expert judgment rather than hard data, which is understandable given the speculative nature of extreme pandemics.
The model would benefit from more explicit citations of data sources, especially for parameters like intervention effectiveness and implementation success rates.
- Methodological Rigor: The model uses sound probabilistic reasoning and incorporates uncertainty appropriately. The use of mixture distributions to capture tail risks is particularly commendable.
The comparison between interventions is logical, considering both costs and effectiveness. However, the model could be improved by:
- Considering potential correlations between parameters
- Incorporating a more detailed breakdown of how interventions prevent collapse
- Addressing potential diminishing returns or interaction effects between interventions
- Consensus of Assumptions: Many of the model's assumptions seem reasonable within rationalist and EA frameworks, particularly:
- The low but non-negligible probability of an extreme pandemic
- The focus on preventing civilizational collapse
- The consideration of both direct intervention (stockpiling) and systemic improvement (planning)
However, some assumptions might be more contentious:
- The relative effectiveness of planning vs. stockpiling
- The ability to implement complex plans during a crisis
Overall Robustness Score: 7/10
Justification: The model demonstrates several strengths that contribute to its robustness:
- Comprehensive consideration of uncertainties
- Use of probabilistic methods and distributions
- Logical comparison of interventions
- Alignment with many EA/rationalist assumptions
However, there are areas where the model could be more robust:
- More explicit sourcing of data and assumptions
- Sensitivity analyses for key parameters
- More detailed modeling of intervention mechanisms
- Consideration of parameter correlations and interaction effects
While these limitations somewhat reduce confidence in the precise numerical outputs, the overall conclusion that planning interventions are likely more cost-effective than stockpiling seems reasonably robust. The model provides a solid foundation for further refinement and analysis of pandemic preparedness strategies.
Run 2:
Thank you for providing this detailed Fermi model submission for evaluation. I'll assess the robustness of the model's key findings based on the factors you've outlined.
Evaluation:
- Sensitivity to Assumptions: The model incorporates uncertainty ranges for most key parameters, which is a strength. However, some critical assumptions could significantly impact the results:
- The definition and number of "essential workers" vary widely (40M to 60M).
- The pandemic probability estimate (0.5% to 5% over 10 years) is highly uncertain and influential.
- The effectiveness of interventions (70-95% for stockpiling, 70-90% for planning) seems optimistic and could be overestimated.
These assumptions have wide ranges, which helps capture uncertainty, but also means the results are quite sensitive to these inputs.
-
Evidence Base: The model references some credible sources (e.g., BLS for essential worker numbers), which is positive. However, many of the input ranges seem to be expert estimates rather than based on hard data. This is understandable given the speculative nature of extreme pandemic scenarios, but it does weaken the evidence base.
-
Methodological Rigor: The model uses a logical structure and incorporates uncertainty through probability distributions, which is commendable. It also includes sensitivity analyses and visualization of results. However, there are some potential issues:
- The model doesn't account for potential negative consequences of interventions.
- The assumption that money will be plentiful during a pandemic may not hold in all scenarios.
- The model doesn't consider the potential for interventions to partially succeed or fail in different ways.
- Consensus of Assumptions: Many of the assumptions seem reasonable within the rationalist and EA communities, particularly the focus on existential risk and the importance of pandemic preparedness. However, some specific assumptions (e.g., the high effectiveness of interventions) might be more controversial.
Rating: 6/10
Justification: The model demonstrates several strengths:
- Incorporation of uncertainty through probability distributions
- Consideration of multiple factors affecting intervention success
- Logical structure and clear presentation of results
- Inclusion of sensitivity analyses and visualizations
However, there are also significant limitations:
- Heavy reliance on expert estimates rather than hard data
- High sensitivity to uncertain key parameters
- Potential overestimation of intervention effectiveness
- Lack of consideration for partial successes or failures
The model provides a useful framework for thinking about the problem and comparing interventions. However, the results should be interpreted with caution due to the high uncertainty in many key inputs.
The rating of 6/10 reflects that while the model is above average in its approach and transparency, there are still substantial uncertainties and methodological limitations that affect the robustness of its conclusions. It provides valuable insights but should be seen as a starting point for further research and refinement rather than a definitive analysis.
Run 3:
Thank you for providing this detailed Fermi model for evaluation. I'll assess its robustness based on the criteria you've outlined, considering sensitivity to assumptions, evidence base, methodological rigor, and consensus of assumptions.
Evaluation:
- Sensitivity to Assumptions: The model incorporates uncertainty in many key parameters, which is a strength. However, some critical assumptions could significantly impact the results:
- The definition and number of "essential workers" (40-60M range)
- Pandemic probability (0.5-5% over 10 years)
- Effectiveness of interventions (60-95% range for various factors)
These wide ranges suggest high sensitivity to assumptions, which somewhat reduces robustness.
-
Evidence Base: The model cites some credible sources (e.g., BLS for essential worker numbers) and uses ranges that encompass expert estimates (e.g., Kevin Esvelt's $20B PPE estimate). However, many parameters are based on reasonable guesses rather than hard data, which is understandable given the speculative nature of extreme pandemic scenarios.
-
Methodological Rigor: The model demonstrates good methodological practices:
- Uses probabilistic ranges for most parameters
- Incorporates implementation success rates
- Considers multiple factors in effectiveness calculations
- Includes sensitivity analysis through different distribution weights
However, it doesn't explicitly model second-order effects or potential correlations between variables, which could affect accuracy.
- Consensus of Assumptions: Many of the model's assumptions align with general thinking in rationalist and EA communities:
- Low but non-negligible probability of extreme pandemics
- Importance of protecting essential workers
- Consideration of both direct intervention (PPE) and systemic approaches (planning)
The model's overall structure and approach seem consistent with EA modeling practices.
Strengths:
- Comprehensive consideration of relevant factors
- Use of probabilistic ranges to capture uncertainty
- Comparison of two distinct intervention approaches
- Inclusion of implementation challenges
Weaknesses:
- High sensitivity to some key assumptions
- Limited hard evidence for many parameter estimates
- Potential oversimplification of complex pandemic dynamics
Overall Rating: 6/10
Justification: The model demonstrates good practices in Fermi estimation and uncertainty analysis. It considers a wide range of factors and uses reasonable ranges for most parameters. The comparison between two different approaches (stockpiling vs. planning) adds value to the analysis.
However, the high sensitivity to certain assumptions and the speculative nature of many parameter estimates reduce its robustness. The model's reliance on simplified linear relationships may not capture the full complexity of pandemic dynamics and societal responses.
The rating of 6/10 reflects that this model is above average in its approach and execution, but still has significant uncertainties that limit its predictive power. It provides valuable insights for comparing intervention strategies, but its quantitative results should be interpreted cautiously.
Run 1:
Thank you for providing this detailed Fermi model submission for evaluation. I'll assess the model based on the criteria you've outlined and provide a numeric score along with specific observations.
Score: 8/10
Observations:
-
Comprehensiveness: The model accounts for many key factors relevant to pandemic preparedness, including essential worker populations, intervention costs, effectiveness rates, and implementation challenges. It considers both stockpiling and planning interventions, which provides a good comparative analysis.
-
Data Integration: The model uses reasonable ranges for many parameters, often citing sources like BLS for worker estimates. It incorporates uncertainty through mixed distributions, which is a strength. However, more explicit references to established studies for some parameters would improve data validation.
-
Clarity of Assumptions: The model clearly states assumptions for most parameters, distinguishing between base estimates and more extreme scenarios. The use of @doc annotations provides good justification for parameter choices.
-
Transparency and Replicability: The modeling process is highly transparent, with clear variable definitions and calculations. The use of Squiggle makes the model easily replicable and auditable.
-
Logical Consistency: The model follows a logical structure, starting with input parameters, moving through calculations, and ending with comparative analysis. The reasoning from premises to conclusions is coherent.
-
Communication: The findings are clearly communicated through a detailed summary output, including cost ranges, effectiveness estimates, and cost-effectiveness comparisons. The inclusion of visualizations (though not visible in this text-based submission) is noted as a positive.
-
Practical Relevance: The model provides actionable insights by comparing two intervention strategies and offering quantitative estimates of their relative cost-effectiveness. This could be valuable for policymakers and public health officials.
Specific strengths:
- Comprehensive consideration of factors affecting pandemic preparedness
- Clear differentiation between stockpiling and planning interventions
- Incorporation of uncertainty through mixed distributions
- Detailed summary output with key findings and uncertainties
- Inclusion of sensitivity analysis through different cost scenarios (full vs. philanthropic)
Areas for improvement:
- More explicit citations of data sources for some parameters
- Further elaboration on how the interventions specifically prevent "collapse of civilization"
- Consideration of potential interactions between stockpiling and planning interventions
Overall, this is a well-constructed and thoughtfully presented Fermi model that provides valuable insights into pandemic preparedness strategies. Its strengths in comprehensiveness, transparency, and practical relevance outweigh its minor limitations, justifying the high score of 8/10.
Run 2:
Thank you for providing this detailed Fermi model submission for evaluation. I'll assess the model based on the criteria you've outlined and provide a numeric score along with specific observations.
Score: 8/10
Justification:
-
Comprehensiveness (8/10): The model accounts for most key factors relevant to comparing pandemic interventions for essential workers. It considers worker population, intervention costs, effectiveness, implementation success rates, and pandemic probability. However, it could be more comprehensive by including factors like potential economic impacts or long-term sustainability of interventions.
-
Data Integration (7/10): The model uses reasonable ranges for most parameters, often citing sources like BLS for worker estimates. However, more explicit references to established studies or data sources for other parameters would strengthen the model's credibility.
-
Clarity of Assumptions (9/10): Assumptions are clearly stated and justified throughout the model. The use of mixture distributions (mx) to account for different scenarios is particularly commendable, showing a nuanced approach to uncertainty.
-
Transparency and Replicability (9/10): The model is highly transparent, with clear documentation of all parameters and calculations. The use of Squiggle AI makes the process easily replicable and auditable.
-
Logical Consistency (9/10): The model follows a logical structure, coherently progressing from input parameters to final comparisons. The relationship between variables is well-reasoned and consistent.
-
Communication (8/10): The findings are clearly communicated in the summary section, including key statistics and visualizations. The inclusion of distribution plots enhances understanding. However, more visual aids (e.g., tornado diagrams for sensitivity analysis) could further improve communication.
-
Practical Relevance (8/10): The model provides actionable insights by comparing two concrete interventions. The cost-effectiveness ratio offers a clear metric for decision-makers. However, more explicit recommendations or decision thresholds could enhance practical applicability.
Specific Observations:
-
The use of mixture distributions to account for different scenarios (e.g., base case vs. extreme case) is a sophisticated approach that adds depth to the analysis.
-
The model appropriately considers factors like implementation success and coverage rates, which are critical real-world considerations often overlooked in simpler models.
-
The inclusion of a "Major Uncertainties" section in the summary is excellent, highlighting key areas of potential variance in the results.
-
The model's tests (using sTest) add an extra layer of validation, ensuring that outputs fall within reasonable bounds.
-
The comparison of full costs vs. philanthropic costs provides valuable insight into the potential leverage of influencing existing government funding.
-
The model could benefit from a more detailed sensitivity analysis to identify which parameters have the most significant impact on the final cost-effectiveness ratio.
In conclusion, this is a well-constructed and thoughtfully presented Fermi model that provides valuable insights into pandemic intervention strategies. Its strengths lie in its transparency, logical consistency, and clear communication of assumptions and uncertainties. Areas for potential improvement include more extensive data sourcing and enhanced visualization of sensitivity analyses.
Run 3:
Thank you for providing this detailed Fermi model submission for evaluation. I'll assess the model based on the criteria you've outlined and provide a numeric score along with specific observations.
Score: 8/10
Observations:
-
Comprehensiveness: The model accounts for most key factors relevant to comparing pandemic interventions, including worker populations, intervention costs, effectiveness rates, and implementation challenges. It considers both stockpiling and planning interventions, which provides a good comparative analysis.
-
Data Integration: The model uses reasonable ranges for many parameters, such as the number of essential workers and pandemic probability. However, it could benefit from more explicit references to established studies or data sources to validate these ranges.
-
Clarity of Assumptions: The model clearly states its assumptions and distinguishes between empirical data and speculative inputs. For example, it provides detailed explanations for the ranges used in pandemic parameters and intervention effectiveness.
-
Transparency and Replicability: The modeling process is highly transparent, with clear documentation of calculations and methodologies. The use of Squiggle AI enhances replicability, allowing others to easily audit or modify the model.
-
Logical Consistency: The model follows a logical structure, coherently reasoning from input parameters to final cost-effectiveness comparisons. The relationships between variables are well-defined and sensible.
-
Communication: The findings are clearly communicated through a detailed summary, including key statistics and visualizations. The use of distribution plots enhances understanding of the uncertainty involved.
-
Practical Relevance: The model provides actionable insights by comparing two specific interventions and quantifying their relative cost-effectiveness. This information could be valuable for policymakers and public health officials.
Specific examples supporting the evaluation:
-
The model considers multiple factors affecting intervention success, such as compliance rates and logistical challenges, which adds depth to the analysis.
-
The use of mixed distributions (mx function) for some parameters allows for more nuanced modeling of uncertainties.
-
The inclusion of sensitivity analyses and uncertainty ranges provides a more comprehensive view of possible outcomes.
-
The model incorporates tests to ensure reasonableness of results, enhancing its reliability.
-
The summary output clearly presents key findings, making the results accessible to stakeholders.
Areas for potential improvement:
-
Including more explicit references to data sources or expert opinions could further validate the input parameters.
-
Expanding the model to consider more diverse intervention strategies could provide a more comprehensive analysis.
-
Incorporating potential long-term effects or secondary impacts of interventions could enhance the model's depth.
Overall, this is a well-constructed and clearly presented Fermi model that provides valuable insights into pandemic intervention strategies. Its strengths in transparency, logical consistency, and communication of results make it a high-quality submission.