HOUSE_OVERSIGHT_016944.jpg

2.33 MB

Extraction Summary

0
People
1
Organizations
0
Locations
0
Events
2
Relationships
4
Quotes

Document Information

Type: Report page / book excerpt
File Size: 2.33 MB
Summary

The text argues for transparency in AI through an "open algorithms" approach, comparing the need for data visibility in AI to that in government accountability. It critiques current machine learning as "brute force" and "dead simple stupid," proposing a next-generation approach that incorporates scientific principles and specific basis functions (like laws of physics or human behavior models) to create more robust systems that require less data.

Organizations (1)

Name Type Context
House Oversight

Relationships (2)

Key Quotes (4)

"If you control the data, then you control the AI."
Source
HOUSE_OVERSIGHT_016944.jpg
Quote #1
"Current AI machine-learning algorithms are, at their core, dead simple stupid."
Source
HOUSE_OVERSIGHT_016944.jpg
Quote #2
"We need trusted data to hold current government to account in terms of what they take in and what they put out, and AI should be no different."
Source
HOUSE_OVERSIGHT_016944.jpg
Quote #3
"The fact that humans have a “commonsense” understanding that they bring to most problems suggests what I call the human strategy"
Source
HOUSE_OVERSIGHT_016944.jpg
Quote #4

Full Extracted Text

Complete text extracted from the document (3,584 characters)

and have the results analyzed by the various stakeholders—rather like elected legislatures were originally intended to do.
If we have the data that go into and out of each decision, we can easily ask, Is this a fair algorithm? Is this AI doing things that we as humans believe are ethical? This human-in-the-loop approach is called “open algorithms;” you get to see what the AIs take as input and what they decide using that input. If you see those two things, you’ll know whether they’re doing the right thing or the wrong thing. It turns out that’s not hard to do. If you control the data, then you control the AI.
One thing people often fail to mention is that all the worries about AI are the same as the worries about today’s government. For most parts of the government—the justice system, et cetera—there’s no reliable data about what they’re doing and in what situation. How can you know whether the courts are fair or not if you don’t know the inputs and the outputs? The same problem arises with AI systems and is addressable in the same way. We need trusted data to hold current government to account in terms of what they take in and what they put out, and AI should be no different.
Next-Generation AI
Current AI machine-learning algorithms are, at their core, dead simple stupid. They work, but they work by brute force, so they need hundreds of millions of samples. They work because you can approximate anything with lots of little simple pieces. That’s a key insight of current AI research—that if you use reinforcement learning for credit-assignment feedback, you can get those little pieces to approximate whatever arbitrary function you want.
But using the wrong functions to make decisions means the AI’s ability to make good decisions won’t generalize. If we give the AI new, different inputs, it may make completely unreasonable decisions. Or if the situation changes, then you need to retrain it. There are amusing techniques to find the “null space” in these AI systems. These are inputs that the AI thinks are valid examples of what it was trained to recognize (e.g., faces, cats, etc.), but to a human they’re crazy examples.
Current AI is doing descriptive statistics in a way that’s not science and would be almost impossible to make into science. To build robust systems, we need to know the science behind data. The systems I view as next-generation AIs result from this science-based approach: If you’re going to create an AI to deal with something physical, then you should build the laws of physics into it as your descriptive functions, in place of those stupid little neurons. For instance, we know that physics uses functions like polynomials, sine waves, and exponentials, so those should be your basis functions and not little linear neurons. By using those more appropriate basis functions, you need a lot less data, you can deal with a lot more noise, and you get much better results.
As in the physics example, if we want to build an AI to work with human behavior, then we need to build the statistical properties of human networks into machine-learning algorithms. When you replace the stupid neurons with ones that capture the basics of human behavior, then you can identify trends with very little data, and you can deal with huge levels of noise.
The fact that humans have a “commonsense” understanding that they bring to most problems suggests what I call the human strategy: Human society is a network just like the neural nets trained for deep learning, but the “neurons” in human society are a lot
141
HOUSE_OVERSIGHT_016944

Discussion 0

Sign in to join the discussion

No comments yet

Be the first to share your thoughts on this epstein document