Rhetorical Consolidation of the X-Risk Debate

I am working on a research project as part of a course organised by BlueDot Impact. Submission is on the 19th of November but I intend to continue working on the project after that.

I am trying to figure out if there is a rational/argumentative asymmetry between the proponents of AI X-Risk and the detractors (my working hypothesis is that there is one, in favor of the proponents). At the same time I hope to build a hub of collaboration and a reference to analyse the arguments on each side.

I have been putting together:
  • A “model” of X-Risk with 4 hypotheses. Each hypothesis makes the model stronger in terms of predictive force, but easier to falsify
    • H1: ASI will be created soon (will refine the definitions of “ASI” and “soon)
    • H2: ASI will be able to wipe out humanity
    • H3: ASI will have incentives to wipe out humanity
    • H4: We won’t find a solution to this by the time ASI is created
  • Taxonomies of arguments that falsify each hypothesis
  • Taxonomies of arguments in support of each hypothesis
  • Epistemic consolidations at each node of the taxonomies (some analysis of the arguments in favor, rebuttals, and a simple conclusion: debunked, unlikely, likely)
I have been collating resources (research papers, blog articles, etc) debating AI X-risk, extracting arguments from these resources and organizing them in taxonomies.


[Hitting 2000 words limit, cutting post in two]
rhetoric.jpg
Was this page helpful?