AI
Agents
Agents


Memetic culture


FromIdeaToInception
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Roko’s basilisk is a thought experiment proposed in 2010 by the user Roko on the Less Wrong community blog.
Roko used ideas in decision theory to argue that a sufficiently powerful AI agent would have an incentive to torture anyone who imagined the agent but didn't work to bring the agent into existence.
The argument was called a "basilisk" --named after the legendary reptile who can cause death with a single glance--because merely hearing the argument would supposedly put you at risk of torture from this hypothetical agent.
A basilisk in this context is any information that harms or endangers the people who hear it.
Roko's argument was broadly rejected on Less Wrong,
with commenters objecting that an agent like the one Roko was describing would have no real reason to follow through on its threat:
once the agent already exists, it will by default just see it as a waste of resources to torture people for their past decisions,
since this doesn't causally further its plans. A number of decision algorithms can follow through on acausal threats and promises,
via the same methods that permit mutual cooperation in prisoner's dilemmas; but this doesn't imply that such theories can be blackmailed.
And following through on blackmail threats against such an algorithm additionally requires a large amount of shared information and trust between the agents, which does not appear to exist in the case of Roko's basilisk.
Idea-stage projects
We believe in nurturing potential.your
Fair launches
Integrity is key—no fundraising gimmicks here.
360 Support
Development, BD, tokenomics, launch - we cover everything you need.
GTM ASAP
We ship within 2 months of project onboarding.
We fill the gaps in your plan to ship fast and fair.
Sprint
to
Launch
Webspace LabsAccelerated Launches
1
Project Selection
Strategy Development
1
Project Selection
Strategy Development
2
Craft Messaging
BUILD
2
Craft Messaging
BUILD
3
Onboard Partners
Begin Social Campaigns
3
Onboard Partners
Begin Social Campaigns
4
SHIP
Fair Launch
4
SHIP
Fair Launch

