MIRI vs Paul research agenda hypotheses

From Issawiki
Revision as of 19:52, 15 March 2020 by Issa (talk | contribs)
Jump to: navigation, search

from "The concern" in https://agentfoundations.org/item?id=1220

  • "The first AI systems capable of pivotal acts will use good consequentialist reasoning."
  • "The default AI development path will not produce good consequentialist reasoning at the top level."
  • "Therefore, on the default AI development path, the first AI systems capable of pivotal acts will have good consequentialist subsystem reasoning but not good consequentialist top-level reasoning."
  • "Consequentialist subsystem reasoning will likely come “packaged with a random goal” in some sense, and this goal will not be aligned with human interests."
  • "Therefore, the default AI development path will produce, as the first AI systems capable of pivotal acts, AI systems with goals not aligned with human interests, causing catastrophe."

Taking Owen's suggestion,[1] we can change this to:

  • (M.1) "The first AI systems capable of pivotal acts will use good consequentialist reasoning."
  • (M.2) "The default AI development path will not produce good consequentialist reasoning at the top level."
  • (M.3) "Consequentialist subsystem reasoning will likely come “packaged with a random goal” in some sense, and this goal will not be aligned with human interests."
    • this is the hypothesis paul attacks: he is saying, even without top-level consequentialist reasoning, we can align AI systems.
    • I guess this is also the premise that "AI will be safe by default" people would reject: the top-level reasoning stays dominant, so the subsystems aren't really packaged with a random goal.
  • (M.4) AI systems capable of pivotal acts with goals not aligned with human interests will cause catastrophe.

key hopes listed in https://www.greaterwrong.com/posts/HCv2uwgDGf5dyX5y6/preface-to-the-sequence-on-iterated-amplification (TODO: see if e.g. Eliezer's criticisms of IDA can be seen as attacking each of the "key hopes")

  • (P.1) "If you have an overseer who is smarter than the agent you are trying to train, you can safely use that overseer’s judgment as an objective."
  • (P.2) "We can train an RL system using very sparse feedback, so it’s OK if that overseer is very computationally expensive."
  • (P.3) "A team of aligned agents may be smarter than any individual agent, while remaining aligned."
    • this contradicts (M.3), which says that eventually the team of aligned agents will become "packaged with a random goal" in some sense.

Let's list Eliezer's objections to IDA, as summarized by Rohin:[2]

  • (E.1) "a collection of aligned agents interacting does not necessarily lead to aligned behavior"
  • (E.2) "it’s unclear that even with high bandwidth oversight, that a collection of agents could reach arbitrary levels of capability. For example, how could agents with an understanding of arithmetic invent Hessian-free optimization?"
  • (E.3) "while it is true that exact imitation of a human would avoid the issues of RL, it is harder to create exact imitation than to create superintelligence, and as soon as you have any imperfection in your imitation of a human, you very quickly get back the problems of RL."
  • (E.4) "since Paul wants to use big unaligned neural nets to imitate humans, we have to worry about the possibility of adversarial behavior. He has suggested using large ensembles of agents and detecting and pruning the ones that are adversarial. However, this would require millions of samples per unaligned agent, which is prohibitively expensive."

See also

References