HOUSE_OVERSIGHT_013123.jpg

2.64 MB

Extraction Summary

7
People
2
Organizations
0
Locations
0
Events
1
Relationships
3
Quotes

Document Information

Type: Academic paper / book chapter (evidentiary production)
File Size: 2.64 MB
Summary

This document is page 207 of an academic text discussing the risks of Artificial General Intelligence (AGI). It reviews concepts such as Friendly AI, Coherent Extrapolated Volition (CEV), and Coherent Blended Volition (CBV), citing researchers like Yudkowsky, Goertzel, and Omohundro. It bears a 'HOUSE_OVERSIGHT_013123' Bates stamp, indicating it was part of a document production to the US House Oversight Committee, likely related to investigations into Jeffrey Epstein's funding of scientific research and connections to academia (e.g., MIT Media Lab).

People (7)

Name Role Context
Kaj Sotala Researcher
Provided objections to Friendly AI concepts.
Stephen Omohundro Researcher
Argued that AI will demonstrate 'basic AI drives' like self-protection and resource acquisition.
Yudkowsky Researcher
Suggested the concept of 'Coherent Extrapolated Volition' (CEV). (Likely Eliezer Yudkowsky).
Bob Jones Hypothetical Subject
Used as an example to explain the difference between desires and aspirations in AI ethics.
Goertzel Researcher
Proposed Coherent Aggregated Volition (CAV). (Likely Ben Goertzel).
Arkin Author/Researcher
Author of 'Governing Lethal Behavior in Autonomous Robots'.
Wallach Researcher
Argued that human values are too complex for compact formal logic rules.

Organizations (2)

Name Type Context
US military
Funded research resulting in Arkin's book on autonomous robots.
House Oversight
Entity identified in the Bates stamp, indicating this document was produced for a congressional investigation.

Relationships (1)

Goertzel Academic/Intellectual Yudkowsky
Goertzel proposed CAV as a related notion to Yudkowsky's CEV.

Key Quotes (3)

"Stephen Omohundro [Omo08] has argued that any advanced AI system will very likely demonstrate certain 'basic AI drives', such as desiring to be rational, to self-protect, to acquire resources"
Source
HOUSE_OVERSIGHT_013123.jpg
Quote #1
"One thought we have recently entertained is that the core concept underlying CAV might be better thought of as CBV or 'Coherent Blended Volition.'"
Source
HOUSE_OVERSIGHT_013123.jpg
Quote #2
"Arkin's excellent though chillingly-titled book Governing Lethal Behavior in Autonomous Robots"
Source
HOUSE_OVERSIGHT_013123.jpg
Quote #3

Full Extracted Text

Complete text extracted from the document (4,242 characters)

12.2 Review of Current Thinking on the Risks of AGI
207
Formalization of human morality has vexed moral philosophers for quite some time. Finally, it is
unclear the extent to which such a proof could be created in a generic, environment-independent
way – but if the proof depends on properties of the physical environment, then it would re-
quire a formalization of the environment itself, which runs up against various problems such
as the complexity of the physical environment and also the fact that we currently have no complete,
consistent theory of physics. Kaj Sotala has provided a list of 14 objections to the Friendly
AI concept, and suggested answers to each of them [Sot11]. Stephen Omohundro [Omo08] has
argued that any advanced AI system will very likely demonstrate certain "basic AI drives", such
as desiring to be rational, to self-protect, to acquire resources, and to preserve and protect its
utility function and avoid counterfeit utility; these drives, he suggests, must be taken carefully
into account in formulating approaches to Friendly AI.
The problem of formally or at least very carefully defining the goal of Friendliness has been
considered from a variety of perspectives, none showing dramatic success. Yudkowsky [Yud04]
has suggested the concept of "Coherent Extrapolated Volition", which roughly refers to the
extrapolation of the common values of the human race. Many subtleties arise in specifying
this concept – e.g. if Bob Jones is often possessed by a strong desire to kill all Martians, but
he deeply aspires to be a nonviolent person, then the CEV approach would not rate "killing
Martians" as part of Bob's contribution to the CEV of humanity.
Goertzel [Goe10a] has proposed a related notion of Coherent Aggregated Volition (CAV),
which eschews the subtleties of extrapolation, and simply seeks a reasonably compact, coherent,
consistent set of values that is fairly close to the collective value-set of humanity. In the CAV
approach, "killing Martians" would be removed from humanity's collective value-set because
it's uncommon and not part of the most compact/coherent/consistent overall model of human
values, rather than because of Bob Jones' aspiration to nonviolence.
One thought we have recently entertained is that the core concept underlying CAV might
be better thought of as CBV or "Coherent Blended Volition." CAV seems to be easily misin-
terpreted as meaning the average of different views, which was not the original intention. The
CBV terminology clarifies that the CBV of a diverse group of people should not be thought of
as an average of their perspectives, but as something more analogous to a "conceptual blend"
[FT02] – incorporating the most essential elements of their divergent views into a whole that is
overall compact, elegant and harmonious. The subtlety here (to which we shall return below)
is that for a CBV blend to be broadly acceptable, the different parties whose views are being
blended must agree to some extent that enough of the essential elements of their own views
have been included. The process of arriving at this sort of consensus may involve extrapolation
of a roughly similar sort to that considered in CEV.
Multiple attempts at axiomatization of human values have also been attempted, e.g. with a
view toward providing near-term guidance to military robots (see e.g. Arkin's excellent though
chillingly-titled book Governing Lethal Behavior in Autonomous Robots [Ark09b], the result
of US military funded research). However, there are reasonably strong arguments that human
values (similarly to e.g. human language or human perceptual classification rules) are too com-
plex and multifaceted to be captured in any compact set of formal logic rules. Wallach [WA10]
has made this point eloquently, and argued the necessity of fusing top-down (e.g. formal logic
based) and bottom-up (e.g. self-organizing learning based) approaches to machine ethics.
A number of more sociological considerations also arise. It is sometimes argued that the risk
from highly-advanced AGI going morally awry on its own may be less than that of moderately-
advanced AGI being used by human beings to advocate immoral ends. This possibility gives
HOUSE_OVERSIGHT_013123

Discussion 0

Sign in to join the discussion

No comments yet

Be the first to share your thoughts on this epstein document