2014: Initial thoughts on AI and what would lead to IdeaFoundry (Genetic/Evolutionary algos)
Muse Notes: An international collaboration between humans and machine
“Keep humankind ultimately in charge of its own destiny”
Evolutionary Algorithms
A population of candidate solutions (which can be data structures or programs) is maintained, and new candidate solutions are generated randomly by mutating or recombining variants in the existing population.
Not generated completely randomly in Avidity
Periodically, the population is pruned by applying a selection criterion (a fitness function) that allows only the better candidates to survive into the next generation.
The fitness function in Avidity is human intuition.
Iterated over thousands of generations, the average quality of the solutions in the candidate pool gradually increases.
The efficiency of Avidity increases
Evolution is computationally demanding and is often defeated by the combinatorial explosion.
Avidity human intuition.
Evolutionary algorithms, however, require not only variations to select among but also a fitness function to evaluate variants, and this is typically the most computationally expensive component.
Genetic Algorithms
General: the Darwinian concept of natural selection as the driving force behind systems for solving real world problems
Bioinformatics:
Steps to AI
A capacity to learn.
Ability to deal effectively with uncertainty and probabilistic information.
Some faculty for extracting useful concepts from sensory data and internal states, and for leveraging acquired concepts into flexible combinatorial representations for use in logical and intuitive reasoning.
Evolutionary processes with foresight— that is, genetic programs designed and guided by an intelligent human programmer.
Computational brute force
Not feasible. Avidity accomplishes with human intuition.
Perhaps we ‘wean’ Muse off of humans.
But always keep humans in the loop
Speed at which the AI grows
Slow: A slow takeoff is one that occurs over some long temporal interval, such as decades or centuries. Slow takeoff scenarios offer excellent opportunities for human political processes to adapt and respond.
Preferable to other scenarios where we don’t have time to react. Implausible but we need this to happen.
The cloud and access to ‘limitless’ computing power in parallel. Need to build our own Muse network with different hardware focused on different in parallel tasks. Not just the option of brute force.
Different algorithms on different scales.
Optimization or efficiency of power
Requires all parts to function correctly. Including humans. Without all parts it cannot function at all
Predictability through design
Even before an agent has been created we might be able to predict something about its behavior, if we know something about who will build it and what goals they will want it to have.
Goal-content integrity
If an agent retains its present goals into the future, then its present goals will be more likely to be achieved by its future self.
What is the final goal? Does it change?
Cognitive Enhancement
Which cognitive abilities are instrumentally useful depends both on the agent’s final goals and on its situation.
Not everything needs full power thrown at it. Some functions more than others.
Speed superintelligence: A system that can do all that a human intellect can do, but much faster.
Humans offload tasks to Muse
Collective superintelligence: A system composed of a large number of smaller intellects such that the system’s overall performance across many very general domains vastly outstrips that of any current cognitive system.
Collective intelligence excels at solving problems that can be readily broken into parts such that solutions to sub-problems can be pursued in parallel and verified independently.
The resulting system would need to be capable of vastly outperforming any current collective intelligence or other cognitive system across many very general domains.
Muse is a collective semi-superintelligence or just a collective intelligence.
Many different parts working together in unison.
“We can think of wisdom as the ability to get the important things approximately right.”
Muse helps take the approximately out of the equation.
Quality superintelligence: A system that is at least as fast as a human mind and vastly qualitatively smarter.
Non-realized cognitive talents, talents that no actual human possesses even though other intelligent systems— ones with no more computing power than the human brain— that did have those talents would gain enormously in their ability to accomplish a wide range of strategically relevant tasks.
What tasks in nature, that humans can’t do, can Muse emulate?
Reinforcement Learning
An area of machine learning that studies techniques whereby agents can learn to maximize some notion of cumulative reward.
What does a program want?
Recommender Systems
Human Intuition + AI
Many cognitive tasks could be performed far more efficiently if the brain’s native support for parallelizable pattern-matching algorithms were complemented by, and integrated with, support for fast sequential processing.
Human collectives are replete with inefficiencies arising from the fact that it is nearly impossible to achieve complete uniformity of purpose among the members of a large group.
Muse and Avidity’s collaborative puzzle solving
Ljjkj
How does Muse communicate?
More of a super organism. Sum of all is parts.
Many humans and many different software components.
AI threats (avoidable by somehow instilling human intuition end empathy, essentially need to make it human)
The treacherous turn
While weak, an AI behaves cooperatively (increasingly so, as it gets smarter). When the AI gets sufficiently strong— without warning or provocation— it strikes, forms a singleton, and begins directly to optimize the world according to the criteria implied by its final values.
Malignant failure modes
Perverse Instantiation
EG: Final goal: “Make us smile” Perverse instantiation: Paralyze human facial musculatures into constant beaming smiles
Mind Crime
Especially difficult to avoid when evolutionary methods are used to produce human-like intelligence , at least if the process is meant to look anything like actual biological evolution.
Managing those threats
Utility function
Decision rule
Capability control methods
Value loading problem
How do we give a machine values?
If the value-loading problem is so tricky, how do we ourselves manage to acquire our values?
Much of the information content in our final values is thus acquired from our experiences rather than preloaded in our genomes.
Motivational scaffolding
Start with small altruistic goals then keep making them more and more complex.
Replace only by group human decision.
Value learning
In contrast to the scaffolding approach, which gives the AI an interim scaffold goal and later replaces it with a different final goal, the value learning approach retains an unchanging final goal throughout the AI’s developmental and operational phases. Learning does not change the goal. It changes only the AI’s beliefs about the goal.
Motivation selection methods
Tripwires
Shuts down after diagnostics find problems
Establish a legal system of sorts
Types of systems
Oracle: Question/Answer system
Keeps it inside the box
Genie: Command (wish) executing system
Sovereign: Open ended anonymous operation
As a Tool: Goal directed behavior
Lie detection
Needs to be an international cooperative effort.
Requires all parts to function correctly. Including humans. Without all parts it cannot function at all
Essentially always keep it hybrid man/machine.
Coherent Extrapolated Volition
Our coherent extrapolated volition is our wish if we knew more, thought faster, were more the people we wished we were, had grown up farther together; where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere; extrapolated as we wish that extrapolated, interpreted as we wish that interpreted.
CEV is meant to be capable of commanding wide support. This is not just because it allocates influence equitably . There is also a deeper ground for the irenic potential of CEV, namely that it enables many different groups to hope that their preferred vision of the future will prevail totally.
What is Morally Right?? (Permissable and Rightness)
Decision Theory
How does the AI make decisions (not related to the human component)?
“Preview AI’s ideas”
As in an oracle.
Needs to be done without any government involvement. An international private consortium. Government and it’s short sightedness/greed/dick waving will just get in the way.
AI as only a human cognitive enhancement.
Again, Collaboration
Collaboration thus offers many benefits. It reduces the haste in developing machine intelligence. It allows for greater investment in safety. It avoids violent conflicts. And it facilitates the sharing of ideas about how to solve the control problem. To these benefits we can add another: collaboration would tend to produce outcomes in which the fruits of a successfully controlled intelligence explosion get distributed more equitably.
Muse Underlying Rules
‘Taught’ starting day one from a multicultural point of view.
Can never make an evolutionary leap without a consensus vote from an all human ‘international board’.



To be clear, credit goes to Nick Bostrom here with a ton of quotes, and with my thoughts on those quotes and ideas.