Difference between revisions of "MIRI vs Paul research agenda hypotheses"

From Issawiki
Jump to: navigation, search
 
(5 intermediate revisions by the same user not shown)
Line 17: Line 17:
 
* (M.4) AI systems capable of pivotal acts with goals not aligned with human interests will cause catastrophe.
 
* (M.4) AI systems capable of pivotal acts with goals not aligned with human interests will cause catastrophe.
  
key hopes listed in https://www.greaterwrong.com/posts/HCv2uwgDGf5dyX5y6/preface-to-the-sequence-on-iterated-amplification (TODO: see if e.g. [[Eliezer]]'s criticisms of [[IDA]] can be seen as attacking each of the "key hopes")
+
key hopes listed in https://www.greaterwrong.com/posts/HCv2uwgDGf5dyX5y6/preface-to-the-sequence-on-iterated-amplification
  
 
* (P.1) "If you have an overseer who is smarter than the agent you are trying to train, you can safely use that overseer’s judgment as an objective."
 
* (P.1) "If you have an overseer who is smarter than the agent you are trying to train, you can safely use that overseer’s judgment as an objective."
 
* (P.2) "We can train an RL system using very sparse feedback, so it’s OK if that overseer is very computationally expensive."
 
* (P.2) "We can train an RL system using very sparse feedback, so it’s OK if that overseer is very computationally expensive."
 
* (P.3) "A team of aligned agents may be smarter than any individual agent, while remaining aligned."
 
* (P.3) "A team of aligned agents may be smarter than any individual agent, while remaining aligned."
 +
** this contradicts (M.3), which says that eventually the team of aligned agents will become "packaged with a random goal" in some sense.
 +
 +
Let's list Eliezer's objections to IDA, as summarized by [[Rohin]]:<ref>https://www.greaterwrong.com/posts/4ap6WQx52txzsJx6r/the-alignment-newsletter-7-05-21-18</ref>
 +
 +
* (E.1) "a collection of aligned agents interacting does not necessarily lead to aligned behavior"
 +
** This directly goes against (P.3), in particular the "while remaining aligned" part.
 +
* (E.2) "it’s unclear that even with high bandwidth oversight, that a collection of agents could reach arbitrary levels of capability. For example, how could agents with an understanding of arithmetic invent Hessian-free optimization?"
 +
** This directly goes against (P.3), in particular the "smarter than any individual agent" part. (Well, to be precise, (P.3) does not actually state that these agents can reach ''arbitrary'' levels of capability, but Paul thinks IDA can compete with unaligned AGI, so the capabilities part would naturally fall under (P.3).)
 +
* (E.3) "while it is true that exact imitation of a human would avoid the issues of RL, it is harder to create exact imitation than to create superintelligence, and as soon as you have any imperfection in your imitation of a human, you very quickly get back the problems of RL."
 +
** This one might not attack one of paul's key hopes, because paul isn't aiming for exact imitation? either that, or the thing it attacks depends on how you cash out "harder to create".
 +
* (E.4) "since Paul wants to use big unaligned neural nets to imitate humans, we have to worry about the possibility of adversarial behavior. He has suggested using large ensembles of agents and detecting and pruning the ones that are adversarial. However, this would require millions of samples per unaligned agent, which is prohibitively expensive."
 +
** This one seems to be attacking (P.2), i.e. it's saying we can't use sparse feedback to train the system.
 +
 +
So it looks like each of [[Eliezer]]'s criticisms of [[IDA]] can mostly be seen as attacking one of the "key hopes" listed by Paul.
  
 
==See also==
 
==See also==
Line 30: Line 44:
  
 
<references/>
 
<references/>
 +
 +
[[Category:AI safety]]

Latest revision as of 03:57, 26 April 2020

from "The concern" in https://agentfoundations.org/item?id=1220

  • "The first AI systems capable of pivotal acts will use good consequentialist reasoning."
  • "The default AI development path will not produce good consequentialist reasoning at the top level."
  • "Therefore, on the default AI development path, the first AI systems capable of pivotal acts will have good consequentialist subsystem reasoning but not good consequentialist top-level reasoning."
  • "Consequentialist subsystem reasoning will likely come “packaged with a random goal” in some sense, and this goal will not be aligned with human interests."
  • "Therefore, the default AI development path will produce, as the first AI systems capable of pivotal acts, AI systems with goals not aligned with human interests, causing catastrophe."

Taking Owen's suggestion,[1] we can change this to:

  • (M.1) "The first AI systems capable of pivotal acts will use good consequentialist reasoning."
  • (M.2) "The default AI development path will not produce good consequentialist reasoning at the top level."
  • (M.3) "Consequentialist subsystem reasoning will likely come “packaged with a random goal” in some sense, and this goal will not be aligned with human interests."
    • this is the hypothesis paul attacks: he is saying, even without top-level consequentialist reasoning, we can align AI systems.
    • I guess this is also the premise that "AI will be safe by default" people would reject: the top-level reasoning stays dominant, so the subsystems aren't really packaged with a random goal.
  • (M.4) AI systems capable of pivotal acts with goals not aligned with human interests will cause catastrophe.

key hopes listed in https://www.greaterwrong.com/posts/HCv2uwgDGf5dyX5y6/preface-to-the-sequence-on-iterated-amplification

  • (P.1) "If you have an overseer who is smarter than the agent you are trying to train, you can safely use that overseer’s judgment as an objective."
  • (P.2) "We can train an RL system using very sparse feedback, so it’s OK if that overseer is very computationally expensive."
  • (P.3) "A team of aligned agents may be smarter than any individual agent, while remaining aligned."
    • this contradicts (M.3), which says that eventually the team of aligned agents will become "packaged with a random goal" in some sense.

Let's list Eliezer's objections to IDA, as summarized by Rohin:[2]

  • (E.1) "a collection of aligned agents interacting does not necessarily lead to aligned behavior"
    • This directly goes against (P.3), in particular the "while remaining aligned" part.
  • (E.2) "it’s unclear that even with high bandwidth oversight, that a collection of agents could reach arbitrary levels of capability. For example, how could agents with an understanding of arithmetic invent Hessian-free optimization?"
    • This directly goes against (P.3), in particular the "smarter than any individual agent" part. (Well, to be precise, (P.3) does not actually state that these agents can reach arbitrary levels of capability, but Paul thinks IDA can compete with unaligned AGI, so the capabilities part would naturally fall under (P.3).)
  • (E.3) "while it is true that exact imitation of a human would avoid the issues of RL, it is harder to create exact imitation than to create superintelligence, and as soon as you have any imperfection in your imitation of a human, you very quickly get back the problems of RL."
    • This one might not attack one of paul's key hopes, because paul isn't aiming for exact imitation? either that, or the thing it attacks depends on how you cash out "harder to create".
  • (E.4) "since Paul wants to use big unaligned neural nets to imitate humans, we have to worry about the possibility of adversarial behavior. He has suggested using large ensembles of agents and detecting and pruning the ones that are adversarial. However, this would require millions of samples per unaligned agent, which is prohibitively expensive."
    • This one seems to be attacking (P.2), i.e. it's saying we can't use sparse feedback to train the system.

So it looks like each of Eliezer's criticisms of IDA can mostly be seen as attacking one of the "key hopes" listed by Paul.

See also

References