<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.issarice.com/index.php?action=history&amp;feed=atom&amp;title=Selection_effect_for_who_builds_AGI</id>
	<title>Selection effect for who builds AGI - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.issarice.com/index.php?action=history&amp;feed=atom&amp;title=Selection_effect_for_who_builds_AGI"/>
	<link rel="alternate" type="text/html" href="https://wiki.issarice.com/index.php?title=Selection_effect_for_who_builds_AGI&amp;action=history"/>
	<updated>2026-04-05T14:33:50Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.31.6</generator>
	<entry>
		<id>https://wiki.issarice.com/index.php?title=Selection_effect_for_who_builds_AGI&amp;diff=1936&amp;oldid=prev</id>
		<title>Issa at 19:18, 27 February 2021</title>
		<link rel="alternate" type="text/html" href="https://wiki.issarice.com/index.php?title=Selection_effect_for_who_builds_AGI&amp;diff=1936&amp;oldid=prev"/>
		<updated>2021-02-27T19:18:12Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 19:18, 27 February 2021&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l1&quot; &gt;Line 1:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 1:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;quot;But also, what I’m more worried about is that the arguments will always be a bit uncertain, and that they will be the kind of arguments that maybe should push a rational person to think that there’s a 20% chance that this will all go wrong, but that some people would just be willing to take such a 20% chance. Or, that they will be selected for being the people… Not everyone will say it’s exactly 20%, some people will say it’s 50%, some people will say it’s 2%. It turns out the ones who think it’s small will be the ones who then unilaterally make these actions, when there are pressures to develop these technologies, economic pressures, or social, or military pressures.&amp;quot; [https://80000hours.org/podcast/episodes/toby-ord-the-precipice-existential-risk-future-humanity/]&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;quot;But also, what I’m more worried about is that the arguments will always be a bit uncertain, and that they will be the kind of arguments that maybe should push a rational person to think that there’s a 20% chance that this will all go wrong, but that some people would just be willing to take such a 20% chance. Or, that they will be selected for being the people… Not everyone will say it’s exactly 20%, some people will say it’s 50%, some people will say it’s 2%. It turns out the ones who think it’s small will be the ones who then unilaterally make these actions, when there are pressures to develop these technologies, economic pressures, or social, or military pressures.&amp;quot; [https://80000hours.org/podcast/episodes/toby-ord-the-precipice-existential-risk-future-humanity/]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt;&amp;#160;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt;&amp;#160;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[Category:AI safety]]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Issa</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.issarice.com/index.php?title=Selection_effect_for_who_builds_AGI&amp;diff=379&amp;oldid=prev</id>
		<title>Issa: Created page with &quot;&quot;But also, what I’m more worried about is that the arguments will always be a bit uncertain, and that they will be the kind of arguments that maybe should push a rational pe...&quot;</title>
		<link rel="alternate" type="text/html" href="https://wiki.issarice.com/index.php?title=Selection_effect_for_who_builds_AGI&amp;diff=379&amp;oldid=prev"/>
		<updated>2020-03-07T22:52:13Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;&amp;quot;But also, what I’m more worried about is that the arguments will always be a bit uncertain, and that they will be the kind of arguments that maybe should push a rational pe...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;&amp;quot;But also, what I’m more worried about is that the arguments will always be a bit uncertain, and that they will be the kind of arguments that maybe should push a rational person to think that there’s a 20% chance that this will all go wrong, but that some people would just be willing to take such a 20% chance. Or, that they will be selected for being the people… Not everyone will say it’s exactly 20%, some people will say it’s 50%, some people will say it’s 2%. It turns out the ones who think it’s small will be the ones who then unilaterally make these actions, when there are pressures to develop these technologies, economic pressures, or social, or military pressures.&amp;quot; [https://80000hours.org/podcast/episodes/toby-ord-the-precipice-existential-risk-future-humanity/]&lt;/div&gt;</summary>
		<author><name>Issa</name></author>
		
	</entry>
</feed>