HOUSE_OVERSIGHT_013114.jpg

2.06 MB

Extraction Summary

2
People
1
Organizations
0
Locations
0
Events
0
Relationships
2
Quotes

Document Information

Type: Scientific/academic paper or book excerpt (page 198)
File Size: 2.06 MB
Summary

This document is page 198 of a scientific or academic text discussing Artificial General Intelligence (AGI) and Cognitive Development. It details heuristics in machine learning regarding the 'Conservation of Number' and explores 'Theory of Mind' using a hypothetical experiment involving a child and DVD boxes. The document bears a 'HOUSE_OVERSIGHT' Bates stamp, suggesting it was part of a document dump related to a congressional investigation, potentially regarding Epstein's funding of scientific research or AI.

People (2)

Name Role Context
Dora the Explorer Fictional Character
Used as an example in a psychological experiment regarding 'Theory of Mind'.
SpongeBob SquarePants Fictional Character
Used as an example in a psychological experiment regarding 'Theory of Mind'.

Organizations (1)

Name Type Context
House Oversight Committee
Appears in the Bates stamp footer (HOUSE_OVERSIGHT_013114), indicating this document is part of a congressional inves...

Key Quotes (2)

"We hypothesize that this problem is hard enough that for an inference-based AGI system to solve it in a developmentally useful way, its inferences must be guided by meta-inferential lessons learned from prior similar problems."
Source
HOUSE_OVERSIGHT_013114.jpg
Quote #1
"A child lacking a theory of mind can not reason through what someone else would think given knowledge other than her own current knowledge."
Source
HOUSE_OVERSIGHT_013114.jpg
Quote #2

Full Extracted Text

Complete text extracted from the document (3,245 characters)

198
11 Stages of Cognitive Development
tomized logical rule-bases to find specialized solutions that solve only this problem fail to fully address the issue, because these solutions don't create knowledge adequate to aid with the solution of related sorts of problems.
We hypothesize that this problem is hard enough that for an inference-based AGI system to solve it in a developmentally useful way, its inferences must be guided by meta-inferential lessons learned from prior similar problems. When approaching a number conservation problem, for example, a reasoning system might draw upon past experience with set-size problems (which may be trial-and-error experience). This is not a simple "machine learning" approach whose scope is restricted to the current problem, but rather a heuristically guided approach which (a) aggregates information from prior experience to guide solution formulation for the problem at hand, and (b) adds the present experience to the set of relevant information about quantification problems for future refinement of thinking.
[DIAGRAM OF BLOCKS]
Fig. 11.6: Conservation of Number
For instance, a very simple context-specific heuristic that a system might learn would be: "When evaluating the truth value of a statement related to the number of objects in a set, it is generally not that useful to explore branches of the backwards-chaining search tree that contain relationships regarding the sizes, masses, or other physical properties of the objects in the set." This heuristic itself may go a long way toward guiding an inference process toward a correct solution to the problem—but it is not something that a mind needs to know "a priori." A concrete-operational stage mind may learn this by data-mining prior instances of inferences involving sizes of sets. Without such experience-based heuristics, the search tree for such a problem will likely be unacceptably large. Even if it is "solvable" without such heuristics, the solutions found may be overly fit to the particular problem and not usefully generalizable.
11.4.2.2 Theory of Mind
Consider this experiment: a preoperational child is shown her favorite "Dora the Explorer" DVD box. Asked what show she's about to see, she'll answer "Dora." However, when her parent plays the disc, it's "SpongeBob SquarePants." If you then ask her what show her friend will expect when given the "Dora" DVD box, she will respond "SpongeBob" although she just answered "Dora" for herself. A child lacking a theory of mind can not reason through what someone else would think given knowledge other than her own current knowledge. Knowledge of self is intrinsically related to the ability to differentiate oneself from others, and this ability may not be fully developed at birth.
Several theorists [BC94, Fod94], based in part on experimental work with autistic children, perceive theory of mind as embodied in an innate module of the mind activated at a certain developmental stage (or not, if damaged). While we consider this possible, we caution against adopting a simplistic view of the "innate vs. acquired" dichotomy: if there is innateness it may take the form of an innate predisposition to certain sorts of learning [EBJ+97].
HOUSE_OVERSIGHT_013114

Discussion 0

Sign in to join the discussion

No comments yet

Be the first to share your thoughts on this epstein document