2 The Alignment Problem


Authors

Last Updated

9/1/2021

ar·ti·fi·cial (source)

adj.

  1. a. Made by humans, especially in imitation of something natural: an artificial sweetener that replaces sugar; artificial flowers.

in·tel·li·gence (source)

n.

  1. The ability to acquire, understand, and use knowledge: a person of extraordinary intelligence.

artificial intelligence The aim of the field of artificial intelligence is to make machines that can acquire, understand, and use knowledge as humans can. It’s not about faking intelligence or about achieving intelligence in a few narrow domains like board games and image recognition. Nor is it about building tools to assist (but not replace) humans. It’s about building machines that can do anything humans can do, and, when possible, do it better.

That puts us in a precarious position. As professor Stuart Russell writes [1]:

The incentives for further development of AI […] are huge, and the momentum appears unstoppable. We must, therefore, ask, “What if we succeed?” This question is seldom considered in the AI literature, which is focused primarily on the pursuit of success rather than on its consequences.

In the face of future adversaries, potentially much more intelligent than we are and with very different goals than we have, those consequences could prove catastrophic.

the alignment problem This is a book about the consequences. It’s about the risks of building machines that are more powerful than we are and about our — still very limited — understanding of how to reduce those risks. It’s about the alignment problem: how do we make sure that the machines we build and deploy pursue our goals and not their own?

We don’t yet know the answer to that question. Given the stakes and the urgency, we can’t afford to wait in writing this book until we do know the answer.

Understanding the Risks


Previous:

Preface

References

Updates & Corrections

If you see any mistakes or want to suggest improvements, please create an issue on GitHub

Reuse

Figures and text are available under a Creative Commons license (CC-BY 4.0) unless otherwise mentioned. The source is available on GitHub. Figures taken from other sources (as mentioned in the captions by "Figure from...") are not available under this license.

Citation

@misc{hoogland2021,
    title={The Alignment Problem},
    author={Jesse Hoogland},
    year={2021},
    url={https://agi-curriculum.vercel.app//0-introduction/1-the-alignment-problem}
}