Daily commit
This commit is contained in:
parent
66e4ac335d
commit
8120cceaf7
|
@ -60908,7 +60908,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 19%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -60929,7 +60929,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/will-macaskill-paralysis-and-hinge-of-history/#transcript",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Will MacAskill (~2019)",
|
||||
"description": "Actual estimate: 1%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61055,7 +61055,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Gott III (~1993)",
|
||||
"description": "Actual estimate: 5%.\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61181,7 +61181,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Global Catastrophic Risk Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61202,7 +61202,7 @@
|
|||
"url": "https://arxiv.org/abs/1705.08807",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Survey of AI experts (~2017)",
|
||||
"description": "Actual estimate: 5%\n\nThe report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/",
|
||||
"description": "The report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61286,7 +61286,7 @@
|
|||
"url": "https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Buck Schlegris (~2020)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61475,7 +61475,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 2%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61601,7 +61601,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61706,7 +61706,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Survey of experts in the AI field (~2016)",
|
||||
"description": "Actual estimate: 18%\n\nThis is the mean. According to Beard et al., the question was \"4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?",
|
||||
"description": "This is the mean. According to Beard et al., the question was \"4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61832,7 +61832,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 10%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61874,7 +61874,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 60%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61895,7 +61895,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 90%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61916,7 +61916,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 97%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61937,7 +61937,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 3%\n\nReduced from his 5% unconditional probability",
|
||||
"description": "Reduced from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -61958,7 +61958,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918905",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 10%\n\nIncreased from his 5% unconditional probability",
|
||||
"description": "Increased from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62000,7 +62000,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918907",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 25%\n\nIncreased from his 5% unconditional probability",
|
||||
"description": "Increased from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62021,7 +62021,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62042,7 +62042,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62168,7 +62168,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 70%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62189,7 +62189,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 67%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62210,7 +62210,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 62%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62231,7 +62231,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 60%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62252,7 +62252,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 52%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62294,7 +62294,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 10%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62315,7 +62315,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62336,7 +62336,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62357,7 +62357,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 60%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62378,7 +62378,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62399,7 +62399,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 25%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62420,7 +62420,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62441,7 +62441,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62462,7 +62462,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 1%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62483,7 +62483,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62504,7 +62504,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62525,7 +62525,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62546,7 +62546,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62630,7 +62630,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62651,7 +62651,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 98%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62672,7 +62672,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62693,7 +62693,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 72%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62714,7 +62714,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 72%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62735,7 +62735,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 70%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62756,7 +62756,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -62777,7 +62777,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 10%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
|
|
@ -13,7 +13,6 @@
|
|||
"title": "Overall risk of extinction prior to 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.19,
|
||||
"actualEstimate": "19%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Total risk",
|
||||
|
@ -23,7 +22,6 @@
|
|||
"title": "Existential risk in the 21st century",
|
||||
"url": "https://80000hours.org/podcast/episodes/will-macaskill-paralysis-and-hinge-of-history/#transcript",
|
||||
"probability": 0.01,
|
||||
"actualEstimate": "1%",
|
||||
"platform": "Will MacAskill",
|
||||
"date_approx": 2019,
|
||||
"category": "Total risk",
|
||||
|
@ -83,7 +81,6 @@
|
|||
"title": "Humanity will cease to exist before 5,100 years or thrive beyond 7.8 million years",
|
||||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%.",
|
||||
"platform": "Gott III",
|
||||
"date_approx": 1993,
|
||||
"category": "Total risk",
|
||||
|
@ -143,7 +140,6 @@
|
|||
"title": "Human extinction by 2100 as a result of superintelligent AI",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "Global Catastrophic Risk Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "AI",
|
||||
|
@ -153,7 +149,6 @@
|
|||
"title": "Extremely bad (e.g. extinction)” long-run impact on humanity from “high-level machine intelligence",
|
||||
"url": "https://arxiv.org/abs/1705.08807",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "Survey of AI experts",
|
||||
"date_approx": 2017,
|
||||
"category": "AI",
|
||||
|
@ -193,7 +188,6 @@
|
|||
"title": "AI-induced existential catastrophe",
|
||||
"url": "https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/",
|
||||
"probability": 0.5,
|
||||
"actualEstimate": "50%",
|
||||
"platform": "Buck Schlegris",
|
||||
"date_approx": 2020,
|
||||
"category": "AI",
|
||||
|
@ -283,7 +277,6 @@
|
|||
"title": "Human extinction by 2100 as a result of single biggest engineered pandemic",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.02,
|
||||
"actualEstimate": "2%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Biorisk",
|
||||
|
@ -343,7 +336,6 @@
|
|||
"title": "Human extinction by 2100 as a result of molecular nanotech weapons",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nanotechnology",
|
||||
|
@ -393,7 +385,6 @@
|
|||
"title": "The probability that the long-run overall impact on humanity of human level machine intelligence will be Extremely bad (existential catastrophe)”, assuming Human Level Machine Intelligence will at some point exist.",
|
||||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"probability": 0.18,
|
||||
"actualEstimate": "18%",
|
||||
"platform": "Survey of experts in the AI field",
|
||||
"date_approx": 2016,
|
||||
"category": "AI/conditional",
|
||||
|
@ -453,7 +444,6 @@
|
|||
"title": "Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 2 degrees of warming",
|
||||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "Mark Lynas",
|
||||
"date_approx": 2020,
|
||||
"category": "Climate change/conditional",
|
||||
|
@ -473,7 +463,6 @@
|
|||
"title": "Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 4 degrees of warming",
|
||||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"probability": 0.6,
|
||||
"actualEstimate": "60%",
|
||||
"platform": "Mark Lynas",
|
||||
"date_approx": 2020,
|
||||
"category": "Climate change/conditional",
|
||||
|
@ -483,7 +472,6 @@
|
|||
"title": "Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 5 degrees of warming",
|
||||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"probability": 0.9,
|
||||
"actualEstimate": "90%",
|
||||
"platform": "Mark Lynas",
|
||||
"date_approx": 2020,
|
||||
"category": "Climate change/conditional",
|
||||
|
@ -493,7 +481,6 @@
|
|||
"title": "Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 6 degrees of warming",
|
||||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"probability": 0.97,
|
||||
"actualEstimate": "97%",
|
||||
"platform": "Mark Lynas",
|
||||
"date_approx": 2020,
|
||||
"category": "Climate change/conditional",
|
||||
|
@ -503,7 +490,6 @@
|
|||
"title": "A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate, but the principle of reproductive freedom prevail[ing]",
|
||||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"probability": 0.03,
|
||||
"actualEstimate": "3%",
|
||||
"platform": "Bryan Caplan",
|
||||
"date_approx": 2006,
|
||||
"category": "Misc/conditional",
|
||||
|
@ -513,7 +499,6 @@
|
|||
"title": "A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate and extensive government regulation",
|
||||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918905",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "Bryan Caplan",
|
||||
"date_approx": 2006,
|
||||
"category": "Misc/conditional",
|
||||
|
@ -533,7 +518,6 @@
|
|||
"title": "A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on the number of independent countries on earth [falling to 1] during the next thousand years",
|
||||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918907",
|
||||
"probability": 0.25,
|
||||
"actualEstimate": "25%",
|
||||
"platform": "Bryan Caplan",
|
||||
"date_approx": 2006,
|
||||
"category": "Misc/conditional",
|
||||
|
@ -543,7 +527,6 @@
|
|||
"title": "At least 1 million dead as a result of superintelligent AI before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -553,7 +536,6 @@
|
|||
"title": "At least 1 billion dead as a result of superintelligent AI before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -613,7 +595,6 @@
|
|||
"title": "Soft AGI takeoff",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.7,
|
||||
"actualEstimate": "70%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -623,7 +604,6 @@
|
|||
"title": "By at least 10 years before human-level AGI is built, debate about AGI risk will be as mainstream as global warming is in 2015",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.67,
|
||||
"actualEstimate": "67%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -633,7 +613,6 @@
|
|||
"title": "A government will build the first human-level AGI, assuming humans build one at all",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.62,
|
||||
"actualEstimate": "62%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -643,7 +622,6 @@
|
|||
"title": "A government will build the first human-level AGI, assuming humans build one at all",
|
||||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"probability": 0.6,
|
||||
"actualEstimate": "60%",
|
||||
"platform": "Pablo Stafforini",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -653,7 +631,6 @@
|
|||
"title": "Human-controlled AGI in expectation would result in less suffering than uncontrolled",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.52,
|
||||
"actualEstimate": "52%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -673,7 +650,6 @@
|
|||
"title": "A design very close to CEV will be implemented in humanity's AGI, conditional on AGI being built (excluding other value-learning approaches and other machine-ethics proposals)",
|
||||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "Pablo Stafforini",
|
||||
"date_approx": 2015,
|
||||
"category": "AI/non-existential",
|
||||
|
@ -683,7 +659,6 @@
|
|||
"title": "At least 1 million dead as a result of the single biggest engineered pandemic before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.3,
|
||||
"actualEstimate": "30%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Biorisk/non-existential",
|
||||
|
@ -693,7 +668,6 @@
|
|||
"title": "At least 1 billion dead as a result of the single biggest engineered pandemic before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Biorisk/non-existential",
|
||||
|
@ -703,7 +677,6 @@
|
|||
"title": "At least 1 million dead as a result of the single biggest natural pandemic before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.6,
|
||||
"actualEstimate": "60%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Biorisk/non-existential",
|
||||
|
@ -713,7 +686,6 @@
|
|||
"title": "At least 1 billion dead as a result of the single biggest natural pandemic before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Biorisk/non-existential",
|
||||
|
@ -723,7 +695,6 @@
|
|||
"title": "At least 1 million dead as a result of molecular nanotech weapons before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.25,
|
||||
"actualEstimate": "25%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nanotechnology/non-existential",
|
||||
|
@ -733,7 +704,6 @@
|
|||
"title": "At least 1 billion dead as a result of molecular nanotech weapons before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nanotechnology/non-existential",
|
||||
|
@ -743,7 +713,6 @@
|
|||
"title": "At least 1 million dead as a result of the single biggest nanotech accident before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.05,
|
||||
"actualEstimate": "5%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nanotechnology/non-existential",
|
||||
|
@ -753,7 +722,6 @@
|
|||
"title": "At least 1 billion dead as a result of the single biggest nanotech accident before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.01,
|
||||
"actualEstimate": "1%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nanotechnology/non-existential",
|
||||
|
@ -763,7 +731,6 @@
|
|||
"title": "At least 1 million dead as a result of all nuclear wars before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.3,
|
||||
"actualEstimate": "30%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nuclear/non-existential",
|
||||
|
@ -773,7 +740,6 @@
|
|||
"title": "At least 1 billion dead as a result of all nuclear wars before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nuclear/non-existential",
|
||||
|
@ -783,7 +749,6 @@
|
|||
"title": "At least 1 million dead as a result of all acts of nuclear terrorism before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.3,
|
||||
"actualEstimate": "30%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nuclear/non-existential",
|
||||
|
@ -793,7 +758,6 @@
|
|||
"title": "At least 1 billion dead as a result of all acts of nuclear terrorism before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Nuclear/non-existential",
|
||||
|
@ -833,7 +797,6 @@
|
|||
"title": "Climate change will cause more suffering than it prevents",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.5,
|
||||
"actualEstimate": "50%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "Climate change/non-existential",
|
||||
|
@ -843,7 +806,6 @@
|
|||
"title": "At least 1 million dead as a result of all wars (including civil wars) before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.98,
|
||||
"actualEstimate": "98%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -853,7 +815,6 @@
|
|||
"title": "At least 1 billion dead as a result of all wars (including civil wars) before 2100",
|
||||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"probability": 0.3,
|
||||
"actualEstimate": "30%",
|
||||
"platform": "GCR Conference",
|
||||
"date_approx": 2008,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -863,7 +824,6 @@
|
|||
"title": "Human-inspired colonization of space will cause more suffering than it prevents if it happens",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.72,
|
||||
"actualEstimate": "72%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -873,7 +833,6 @@
|
|||
"title": "Earth will eventually be controlled by a singleton of some sort",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.72,
|
||||
"actualEstimate": "72%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -883,7 +842,6 @@
|
|||
"title": "Earth will eventually be controlled by a singleton of some sort",
|
||||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"probability": 0.7,
|
||||
"actualEstimate": "70%",
|
||||
"platform": "Pablo Stafforini",
|
||||
"date_approx": 2015,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -893,7 +851,6 @@
|
|||
"title": "Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)",
|
||||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"probability": 0.5,
|
||||
"actualEstimate": "50%",
|
||||
"platform": "Brian Tomasik",
|
||||
"date_approx": 2015,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
@ -903,7 +860,6 @@
|
|||
"title": "Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)",
|
||||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"probability": 0.1,
|
||||
"actualEstimate": "10%",
|
||||
"platform": "Pablo Stafforini",
|
||||
"date_approx": 2015,
|
||||
"category": "Miscellaneous/non-existential",
|
||||
|
|
|
@ -4,12 +4,8 @@
|
|||
Ord writes: ""Don’t take these numbers to be completely objective. [...] And don’t take the estimates to be precise. Their purpose is to show the right order of magnitude, rather than a more precise probability.""
|
||||
|
||||
This estimate already incorporates Ord's expectation that people will start taking these risks more seriously in future. For his ""business as usual"" estimate, see the conditional estimates sheet.","[{""name"":""Yes"",""probability"":0.17,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.83,""type"":""PROBABILITY""}]",,,2
|
||||
"Overall risk of extinction prior to 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 19%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.19,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.81,""type"":""PROBABILITY""}]",,,2
|
||||
"Existential risk in the 21st century","https://80000hours.org/podcast/episodes/will-macaskill-paralysis-and-hinge-of-history/#transcript","X-risk estimates","Actual estimate: 1%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.01,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.99,""type"":""PROBABILITY""}]",,,2
|
||||
"Overall risk of extinction prior to 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.19,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.81,""type"":""PROBABILITY""}]",,,2
|
||||
"Existential risk in the 21st century","https://80000hours.org/podcast/episodes/will-macaskill-paralysis-and-hinge-of-history/#transcript","X-risk estimates","","[{""name"":""Yes"",""probability"":0.01,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.99,""type"":""PROBABILITY""}]",,,2
|
||||
"Extinction risk in the next century","https://80000hours.org/articles/extinction-risk/","X-risk estimates","Actual estimate: Probably at or above 3%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.03,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.97,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -25,9 +21,7 @@ That median and mean is as of 3rd July 2019.","[{""name"":""Yes"",""probability"
|
|||
"Existential disaster will do us in","https://www.nickbostrom.com/existential/risks.html","X-risk estimates","Actual estimate: Probably at or above 25%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.25,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.75,""type"":""PROBABILITY""}]",,,2
|
||||
"Humanity will cease to exist before 5,100 years or thrive beyond 7.8 million years","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","Actual estimate: 5%.
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Humanity will cease to exist before 5,100 years or thrive beyond 7.8 million years","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Annual probability as of 2009 of extinction","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","Actual estimate: 0.3-0.4%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.0035,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9965,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -43,12 +37,8 @@ I think it's fairly likely(>20%) that sentient life will survive for at least bi
|
|||
"Existential catastrophe by 2120 as a result of unaligned AI","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","Actual estimate: ~10%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of superintelligent AI","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Extremely bad (e.g. extinction)” long-run impact on humanity from “high-level machine intelligence","https://arxiv.org/abs/1705.08807","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
The report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of superintelligent AI","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Extremely bad (e.g. extinction)” long-run impact on humanity from “high-level machine intelligence","https://arxiv.org/abs/1705.08807","X-risk estimates","The report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"A state where civilization collapses and does not recover, or a situation where all human life ends, due to AI","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","Actual estimate: 0-10%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -58,9 +48,7 @@ Garfinkel was asked for his estimate during an AMA, and replied ""I currently gi
|
|||
"Chance that AI, through adversarial optimization against humans only, will cause existential catastrophe","https://www.lesswrong.com/posts/TdwpN484eTbPSvZkm/rohin-shah-on-reasons-for-ai-optimism","X-risk estimates","Actual estimate: ~5%
|
||||
|
||||
This is my interpretation of some comments that may not have been meant to be taken very literally. Elsewhere, Rohin noted that this was “[his] opinion before updating on other people's views"": https://forum.effectivealtruism.org/posts/tugs9KQyNqi4yRTsb/does-80-000-hours-focus-too-much-on-ai-risk#ZmtPji3pQaZK7Y4FF I think he updated this in 2020 to ~9%, due to pessimism about discontinuous scenarios: https://www.lesswrong.com/posts/TdwpN484eTbPSvZkm/rohin-shah-on-reasons-for-ai-optimism?commentId=n577gwGB3vRpwkBmj Rohin also discusses his estimates here: https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"AI-induced existential catastrophe","https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/","X-risk estimates","Actual estimate: 50%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"AI-induced existential catastrophe","https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"Existential risk from unaligned AI over the coming 100 years","https://forum.effectivealtruism.org/posts/2sMR7n32FSvLCoJLQ/critical-review-of-the-precipice-a-reassessment-of-the-risks","X-risk estimates","Actual estimate: 0.05%
|
||||
|
||||
This was a direct response to Ord's estimate. It focuses on one pathway to x-risk from AI, not all pathways (e.g., not AI misuse or risks from competition between powerful AIs). ""These estimates should not be taken very seriously. I do not believe we have enough information to make sensible quantitative estimates about these eventualities. Nevertheless, I present my estimates largely in order to illustrate the extent of my disagreement with Ord’s estimates, and to illustrate the key considerations I examine in order to arrive at an estimate."" In comments on the source, Rohin Shah critiques some of the inputs to this estimate, and provides his own, substantially higher estimates.","[{""name"":""Yes"",""probability"":0.0005,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9995,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -85,9 +73,7 @@ This is the median. Beard et al.'s appendix says ""Note that for these predictio
|
|||
"Existential catastrophe from naturally arising pandemics by 2120","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0","X-risk estimates","Actual estimate: ~0.01% (~1 in 10,000)
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.0001,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9999,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of single biggest engineered pandemic","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 2%
|
||||
|
||||
This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.02,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.98,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of single biggest engineered pandemic","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.02,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.98,""type"":""PROBABILITY""}]",,,2
|
||||
"Annual probability of an existential catastrophe arising from a global pandemic","https://www.liebertpub.com/doi/10.1089/hs.2017.0028","X-risk estimates","Actual estimate: 0.008% to 0.0000016% (between 8 x 10-5 and 1.6 x 10-8)
|
||||
|
||||
The fact that there's a separate estimate from the same source for biowarfare and bioterrorism suggests to me that this is meant to be an estimate of the risk from a natural pandemic only. But I'm not sure. This might also include ""accidental"" release of a bioengineered pathogen.","[{""name"":""Yes"",""probability"":0.00004,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.99996,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -103,9 +89,7 @@ The fact that there's a separate estimate from the same source for ""synthetic b
|
|||
"Extinction risk from engineered pandemics over the coming 100 years","https://forum.effectivealtruism.org/posts/2sMR7n32FSvLCoJLQ/critical-review-of-the-precipice-a-reassessment-of-the-risks","X-risk estimates","Actual estimate: 0.0002%
|
||||
|
||||
This was a direct response to Ord's estimate, although this estimate is of extinction risk rather than existential risk. ""These estimates should not be taken very seriously. I do not believe we have enough information to make sensible quantitative estimates about these eventualities. Nevertheless, I present my estimates largely in order to illustrate the extent of my disagreement with Ord’s estimates, and to illustrate the key considerations I examine in order to arrive at an estimate."" In comments on the source, Will Bradshaw critiques some of the inputs to this estimate.","[{""name"":""Yes"",""probability"":0.000002,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.999998,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of molecular nanotech weapons","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of molecular nanotech weapons","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"Human extinction by 2100 as a result of the single biggest nanotech accident","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 0.5%
|
||||
|
||||
This is the median. Beard et al.'s appendix says ""Note that for these predictions no time frame was given."" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.","[{""name"":""Yes"",""probability"":0.005,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.995,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -124,9 +108,7 @@ My best guess for that is actually about one in three this century. If we carry
|
|||
Arden Koehler replies """"Okay. So business as usual means doing what we are approximately doing now extrapolated into the future but we don’t put much more effort into it as opposed to doing nothing at all?""""
|
||||
|
||||
Ord replies: """"That’s right, and it turns out to be quite hard to define business as usual. That’s the reason why, for my key estimate, that I make it… In some sense, it’s difficult to define estimates where they take into account whether or not people follow the advice that you’re giving; that introduces its own challenges. But at least that’s just what a probability normally means. It means that your best guess of the chance something happens, whereas a best guess that something happens conditional upon certain trends either staying at the same level or continuing on the same trajectory or something is just quite a bit more unclear as to what you’re even talking about.""""","[{""name"":""Yes"",""probability"":0.33,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.6699999999999999,""type"":""PROBABILITY""}]",,,2
|
||||
"The probability that the long-run overall impact on humanity of human level machine intelligence will be Extremely bad (existential catastrophe)”, assuming Human Level Machine Intelligence will at some point exist.","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904","X-risk estimates","Actual estimate: 18%
|
||||
|
||||
This is the mean. According to Beard et al., the question was ""4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?","[{""name"":""Yes"",""probability"":0.18,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.8200000000000001,""type"":""PROBABILITY""}]",,,2
|
||||
"The probability that the long-run overall impact on humanity of human level machine intelligence will be Extremely bad (existential catastrophe)”, assuming Human Level Machine Intelligence will at some point exist.","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904","X-risk estimates","This is the mean. According to Beard et al., the question was ""4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?","[{""name"":""Yes"",""probability"":0.18,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.8200000000000001,""type"":""PROBABILITY""}]",,,2
|
||||
"Chance that AI, through “adversarial optimization against humans only”, will cause existential catastrophe, conditional on there not being “additional intervention by longtermists” (or perhaps “no intervention from longtermists”)","https://www.lesswrong.com/posts/TdwpN484eTbPSvZkm/rohin-shah-on-reasons-for-ai-optimism","X-risk estimates","Actual estimate: ~10%
|
||||
|
||||
This is my interpretation of some comments that may not have been meant to be taken very literally. I think he updated this in 2020 to ~15%, due to pessimism about discontinuous scenarios: https://www.lesswrong.com/posts/TdwpN484eTbPSvZkm/rohin-shah-on-reasons-for-ai-optimism?commentId=n577gwGB3vRpwkBmj Rohin also discusses his estimates here: https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -145,44 +127,26 @@ I give existential risk over the next century from nuclear war at about one in a
|
|||
"Global human population of zero resulting from the 150 Tg of black carbon scenario in our 2007 paper","http://www.overcomingbias.com/2012/11/nuclear-winter-and-human-extinction-qa-with-luke-oman.html","X-risk estimates","Actual estimate: 0.001-0.01% (“in the range of 1 in 10,000 to 1 in 100,000”)
|
||||
|
||||
I think that this is Oman’s estimate of the chance that extinction would occur if that black carbon scenario occurred, rather than an estimate that also takes into account the low probability that that black carbon scenario occurs. I.e., I think that this estimate was conditional on a particular type of nuclear war occurring. But I’m not sure about that, and the full context doesn’t make it much clearer.","[{""name"":""Yes"",""probability"":0.000055,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.999945,""type"":""PROBABILITY""}]",,,2
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 2 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 2 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
Mark Lynas: ""Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 3 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Actual estimate: 30-40%
|
||||
|
||||
Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
Mark Lynas: ""Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.","[{""name"":""Yes"",""probability"":0.35,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.65,""type"":""PROBABILITY""}]",,,2
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 4 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Actual estimate: 60%
|
||||
|
||||
Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 4 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
Mark Lynas: ""Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.","[{""name"":""Yes"",""probability"":0.6,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.4,""type"":""PROBABILITY""}]",,,2
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 5 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Actual estimate: 90%
|
||||
|
||||
Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 5 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
Mark Lynas: ""Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.","[{""name"":""Yes"",""probability"":0.9,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.09999999999999998,""type"":""PROBABILITY""}]",,,2
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 6 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Actual estimate: 97%
|
||||
|
||||
Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
"Full-scale collapse of society, perhaps due to very, very widespread famine, if there's 6 degrees of warming","https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/","X-risk estimates","Arden Koehler: ""...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?
|
||||
Mark Lynas: ""Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.","[{""name"":""Yes"",""probability"":0.97,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.030000000000000027,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate, but the principle of reproductive freedom prevail[ing]","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904","X-risk estimates","Actual estimate: 3%
|
||||
|
||||
Reduced from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.03,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.97,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate and extensive government regulation","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918905","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
Increased from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate, but the principle of reproductive freedom prevail[ing]","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904","X-risk estimates","Reduced from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.03,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.97,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on genetic screening for personality traits becom[ing] cheap and accurate and extensive government regulation","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918905","X-risk estimates","Increased from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on the number of independent countries on earth [not decreasing] during the next thousand years","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918906","X-risk estimates","Actual estimate: 0.1%
|
||||
|
||||
Reduced from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.001,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.999,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on the number of independent countries on earth [falling to 1] during the next thousand years","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918907","X-risk estimates","Actual estimate: 25%
|
||||
|
||||
Increased from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.25,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.75,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of superintelligent AI before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of superintelligent AI before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"A world totalitarian government will emerge during the next one thousand years and last for a thousand years or more, conditional on the number of independent countries on earth [falling to 1] during the next thousand years","https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918907","X-risk estimates","Increased from his 5% unconditional probability","[{""name"":""Yes"",""probability"":0.25,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.75,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of superintelligent AI before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of superintelligent AI before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"AI safety is as hard as a (caricature of) MIRI suggests","https://aiimpacts.org/conversation-with-adam-gleave/","X-risk estimates","Actual estimate: ~10%
|
||||
|
||||
So, decent chance– I think I put a reasonable probability, like 10% probability, on the hard-mode MIRI version of the world being true. In which case, I think there’s probably nothing we can do.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -200,63 +164,27 @@ I haven’t actually written down these numbers since I last changed my mind abo
|
|||
Basically, you can look at my [estimate that the existential risk from AI in the next 100 years is] 10% as, there’s about a 50% chance that we create something that’s more intelligent than humanity this century. And then there’s only an 80% chance that we manage to survive that transition, being in charge of our future. If you put that together, you get a 10% chance that’s the time where we lost control of the future in a negative way.
|
||||
|
||||
Toby Ord: With that number, I’ve spent a lot of time thinking about this. Actually, my first degree was in computer science, and I’ve been involved in artificial intelligence for a long time, although it’s not what I did my PhD on. But, if you ask the typical AI expert’s view of the chance that we develop smarter than human AGI, artificial general intelligence, this century is about 50%. If you survey the public, which has been done, it’s about 50%. So, my 50% is both based on the information I know actually about what’s going on in AI, and also is in line with all of the relevant outside views. It feels difficult to have a wildly different number on that. The onus would be on the other person.","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"Soft AGI takeoff","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 70%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.7,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.30000000000000004,""type"":""PROBABILITY""}]",,,2
|
||||
"By at least 10 years before human-level AGI is built, debate about AGI risk will be as mainstream as global warming is in 2015","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 67%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.67,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.32999999999999996,""type"":""PROBABILITY""}]",,,2
|
||||
"A government will build the first human-level AGI, assuming humans build one at all","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 62%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.62,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.38,""type"":""PROBABILITY""}]",,,2
|
||||
"A government will build the first human-level AGI, assuming humans build one at all","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","Actual estimate: 60%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.6,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.4,""type"":""PROBABILITY""}]",,,2
|
||||
"Human-controlled AGI in expectation would result in less suffering than uncontrolled","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 52%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.52,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.48,""type"":""PROBABILITY""}]",,,2
|
||||
"Soft AGI takeoff","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.7,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.30000000000000004,""type"":""PROBABILITY""}]",,,2
|
||||
"By at least 10 years before human-level AGI is built, debate about AGI risk will be as mainstream as global warming is in 2015","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.67,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.32999999999999996,""type"":""PROBABILITY""}]",,,2
|
||||
"A government will build the first human-level AGI, assuming humans build one at all","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.62,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.38,""type"":""PROBABILITY""}]",,,2
|
||||
"A government will build the first human-level AGI, assuming humans build one at all","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.6,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.4,""type"":""PROBABILITY""}]",,,2
|
||||
"Human-controlled AGI in expectation would result in less suffering than uncontrolled","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.52,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.48,""type"":""PROBABILITY""}]",,,2
|
||||
"A design very close to CEV will be implemented in humanity's AGI, conditional on AGI being built (excluding other value-learning approaches and other machine-ethics proposals)","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 0.5%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.005,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.995,""type"":""PROBABILITY""}]",,,2
|
||||
"A design very close to CEV will be implemented in humanity's AGI, conditional on AGI being built (excluding other value-learning approaches and other machine-ethics proposals)","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest engineered pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 30%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest engineered pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest natural pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 60%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.6,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.4,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest natural pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of molecular nanotech weapons before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 25%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.25,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.75,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of molecular nanotech weapons before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest nanotech accident before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 5%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest nanotech accident before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 1%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.01,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.99,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all nuclear wars before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 30%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all nuclear wars before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all acts of nuclear terrorism before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 30%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all acts of nuclear terrorism before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"A design very close to CEV will be implemented in humanity's AGI, conditional on AGI being built (excluding other value-learning approaches and other machine-ethics proposals)","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest engineered pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest engineered pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest natural pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.6,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.4,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest natural pandemic before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of molecular nanotech weapons before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.25,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.75,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of molecular nanotech weapons before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of the single biggest nanotech accident before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of the single biggest nanotech accident before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.01,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.99,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all nuclear wars before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all nuclear wars before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all acts of nuclear terrorism before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all acts of nuclear terrorism before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"chance of a full-scale nuclear war in the next century","https://80000hours.org/podcast/episodes/toby-ord-the-precipice-existential-risk-future-humanity/#transcript","X-risk estimates","Actual estimate: ~5%
|
||||
|
||||
I give existential risk over the next century from nuclear war at about one in a thousand. I initially thought it would be higher than that. That’s actually something that while researching the book, thought was a lower risk than I had initially thought. And how I’d break it down is to something like a 5% chance of a full-scale nuclear war in the next century and a 2% chance that that would be the end of human potential."" Ord discusses his reasoning more both in that interview and in The Precipice.","[{""name"":""Yes"",""probability"":0.05,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.95,""type"":""PROBABILITY""}]",,,2
|
||||
|
@ -266,27 +194,11 @@ In this post, I get a rough sense of how probable a nuclear war might be by look
|
|||
"Per year chance of nuclear war between the US and Russia","https://forum.effectivealtruism.org/posts/PAYa6on5gJKwAywrF/how-likely-is-a-nuclear-exchange-between-the-us-and-russia-1","X-risk estimates","Actual estimate: 0.38%
|
||||
|
||||
In this post, I get a rough sense of how probable a nuclear war might be by looking at historical evidence, the views of experts, and predictions made by forecasters. I find that, if we aggregate those perspectives, there’s about a 1.1% chance of nuclear war each year, and that the chances of a nuclear war between the US and Russia, in particular, are around 0.38% per year."" This is not presented as Luisa's own credence; this may not be the number she herself would give. Readers may also be interested in the estimates implied by each of the perspectives Luisa aggregates.","[{""name"":""Yes"",""probability"":0.0038,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9962,""type"":""PROBABILITY""}]",,,2
|
||||
"Climate change will cause more suffering than it prevents","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 50%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all wars (including civil wars) before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 98%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.98,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.020000000000000018,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all wars (including civil wars) before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","Actual estimate: 30%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"Human-inspired colonization of space will cause more suffering than it prevents if it happens","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 72%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.72,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.28,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth will eventually be controlled by a singleton of some sort","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 72%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.72,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.28,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth will eventually be controlled by a singleton of some sort","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","Actual estimate: 70%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.7,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.30000000000000004,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","Actual estimate: 50%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","Actual estimate: 10%
|
||||
|
||||
","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
||||
"Climate change will cause more suffering than it prevents","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 million dead as a result of all wars (including civil wars) before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","","[{""name"":""Yes"",""probability"":0.98,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.020000000000000018,""type"":""PROBABILITY""}]",,,2
|
||||
"At least 1 billion dead as a result of all wars (including civil wars) before 2100","https://www.fhi.ox.ac.uk/reports/2008-1.pdf","X-risk estimates","","[{""name"":""Yes"",""probability"":0.3,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.7,""type"":""PROBABILITY""}]",,,2
|
||||
"Human-inspired colonization of space will cause more suffering than it prevents if it happens","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.72,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.28,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth will eventually be controlled by a singleton of some sort","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.72,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.28,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth will eventually be controlled by a singleton of some sort","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.7,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.30000000000000004,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)","https://reducing-suffering.org/summary-beliefs-values-big-questions/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.5,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.5,""type"":""PROBABILITY""}]",,,2
|
||||
"Earth-originating intelligence will colonize the entire galaxy (ignoring anthropic arguments)","http://www.stafforini.com/blog/what_i_believe/","X-risk estimates","","[{""name"":""Yes"",""probability"":0.1,""type"":""PROBABILITY""},{""name"":""No"",""probability"":0.9,""type"":""PROBABILITY""}]",,,2
|
|
|
@ -24,7 +24,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 19%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -44,7 +44,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/will-macaskill-paralysis-and-hinge-of-history/#transcript",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Will MacAskill (~2019)",
|
||||
"description": "Actual estimate: 1%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -164,7 +164,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=0",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Gott III (~1993)",
|
||||
"description": "Actual estimate: 5%.\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -284,7 +284,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Global Catastrophic Risk Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -304,7 +304,7 @@
|
|||
"url": "https://arxiv.org/abs/1705.08807",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Survey of AI experts (~2017)",
|
||||
"description": "Actual estimate: 5%\n\nThe report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/",
|
||||
"description": "The report's authors discuss potential concerns around non-response bias and the fact that “NIPS and ICML authors are representative of machine learning but not of the field of artificial intelligence as a whole”. There was also evidence of apparent inconsistencies in estimates of AI timelines as a result of small changes to how questions were asked, providing further reason to wonder how meaningful these experts’ predictions were. https://web.archive.org/web/20171030220008/https://aiimpacts.org/some-survey-results/",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -384,7 +384,7 @@
|
|||
"url": "https://futureoflife.org/2020/04/15/an-overview-of-technical-ai-alignment-in-2018-and-2019-with-buck-shlegeris-and-rohin-shah/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Buck Schlegris (~2020)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -564,7 +564,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 2%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -684,7 +684,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"description": "This is the median. Beard et al.'s appendix says \"Note that for these predictions no time frame was given.\" I think that that's incorrect, based on phrasings in the original source, but I'm not certain.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -784,7 +784,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Survey of experts in the AI field (~2016)",
|
||||
"description": "Actual estimate: 18%\n\nThis is the mean. According to Beard et al., the question was \"4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?",
|
||||
"description": "This is the mean. According to Beard et al., the question was \"4. Assume for the purpose of this question that such Human Level Machine Intelligence (HLMI) will at some point exist. How positive or negative would be overall impact on humanity, in the long run?",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -904,7 +904,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 10%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -944,7 +944,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 60%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -964,7 +964,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 90%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -984,7 +984,7 @@
|
|||
"url": "https://80000hours.org/podcast/episodes/mark-lynas-climate-change-nuclear-energy/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Mark Lynas (~2020)",
|
||||
"description": "Actual estimate: 97%\n\nArden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"description": "Arden Koehler: \"...do you have a guess at what degree of warming we would need to reach for the full-scale collapse of society, perhaps due to very, very widespread famine to have say a 10% chance of happening?\nMark Lynas: \"Oh, I think… You want to put me on the spot. I would say it has a 30 to 40% chance of happening at three degrees, and a 60% chance of happening at four degrees, and 90% at five degrees, and 97% at six degrees. [...] Maybe 10% at two degrees.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1004,7 +1004,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918904",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 3%\n\nReduced from his 5% unconditional probability",
|
||||
"description": "Reduced from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1024,7 +1024,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918905",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 10%\n\nIncreased from his 5% unconditional probability",
|
||||
"description": "Increased from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1064,7 +1064,7 @@
|
|||
"url": "https://docs.google.com/spreadsheets/d/1W10B6NJjicD8O0STPiT3tNV3oFnT8YsfjmtYR8RO_RI/edit#gid=511918907",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Bryan Caplan (~2006)",
|
||||
"description": "Actual estimate: 25%\n\nIncreased from his 5% unconditional probability",
|
||||
"description": "Increased from his 5% unconditional probability",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1084,7 +1084,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1104,7 +1104,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A. Interestingly, this is the same as the estimate from this source of the chance of human as a result of superintelligent AI by 2100.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1224,7 +1224,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 70%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1244,7 +1244,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 67%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1264,7 +1264,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 62%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1284,7 +1284,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 60%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1304,7 +1304,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 52%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1344,7 +1344,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 10%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1364,7 +1364,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1384,7 +1384,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1404,7 +1404,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 60%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1424,7 +1424,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1444,7 +1444,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 25%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1464,7 +1464,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1484,7 +1484,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 5%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1504,7 +1504,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 1%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1524,7 +1524,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1544,7 +1544,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1564,7 +1564,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1584,7 +1584,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 10%\n\nThis is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"description": "This is the median. The report about these estimates also plots the results for each question “with individual response distributions visible” in Appendix A.",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1664,7 +1664,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1684,7 +1684,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 98%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1704,7 +1704,7 @@
|
|||
"url": "https://www.fhi.ox.ac.uk/reports/2008-1.pdf",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "GCR Conference (~2008)",
|
||||
"description": "Actual estimate: 30%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1724,7 +1724,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 72%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1744,7 +1744,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 72%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1764,7 +1764,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 70%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1784,7 +1784,7 @@
|
|||
"url": "https://reducing-suffering.org/summary-beliefs-values-big-questions/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Brian Tomasik (~2015)",
|
||||
"description": "Actual estimate: 50%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
@ -1804,7 +1804,7 @@
|
|||
"url": "http://www.stafforini.com/blog/what_i_believe/",
|
||||
"platform": "X-risk estimates",
|
||||
"author": "Pablo Stafforini (~2015)",
|
||||
"description": "Actual estimate: 10%\n\n",
|
||||
"description": "",
|
||||
"options": [
|
||||
{
|
||||
"name": "Yes",
|
||||
|
|
|
@ -11,9 +11,9 @@ let data = JSON.parse(rawdata)
|
|||
let results = []
|
||||
for(let datum of data){
|
||||
let probability = datum["probability"]
|
||||
let description = `Actual estimate: ${datum["actualEstimate"]}
|
||||
let description = datum["actualEstimate"] ? `Actual estimate: ${datum["actualEstimate"]}
|
||||
|
||||
${datum["description"]}`
|
||||
${datum["description"]}` : datum["description"]
|
||||
let author = `${datum["platform"]} (~${datum["date_approx"]})`
|
||||
let result = ({
|
||||
"title": datum["title"],
|
||||
|
|
Loading…
Reference in New Issue
Block a user