Questions for ‘How to design artificial intelligence that acts nice — and only nice’
To accompany ‘How to design artificial intelligence that acts nice — and only nice’
SCIENCE
Before Reading:
- Give one example of a movie, book or series involving robots or artificial intelligence (AI) gone wrong. Describe the plot in one sentence. Describe one aspect of this fictional work that you consider far-fetched, or very unlikely to ever occur in the real world. Consider the recent advancements in AI technology, then describe some potential problem — perhaps from this movie/book/series — that you think could become a reality if we do not take preventative steps.
- Think about the way you’ve used AI technology lately. Write down one ability carried out by AI that improves your life. Then, write down one ability you think an AI of the future should not have, even if it’s technically possible to grant this ability.
During Reading:
- When scientists in this story talk about solving the “alignment problem,” what goal are they trying to achieve?
- To what extent does the Center for AI Safety consider advanced AI to be a serious potential threat to humans? Explain your answer.
- List the “three H’s.”
- Give one example of a large language model.
- How do large language models make use of probabilities to write text?
- How did OpenAI teach its large language model what types of text it should not generate?
- What was OpenAI’s Minecraft bot attempting to accomplish when it chopped down a house?
- Briefly describe the bowl task used to test the AI system called KnowNo.
- This story describes many examples of how scientists and professional researchers help improve AI safety. But how might non-scientists contribute to this goal? Give one example.
After Reading:
- Jeff Hawkins says the fear of AI spontaneously taking over the world is “far-fetched.” Briefly explain his reasoning. Hawkins points out that humans give AI their motivations. In other words, an AI model will want to accomplish whatever it’s told to accomplish. Assuming it’s true that AI will not create its own motivations, how might problems still arise? Describe one approach from this story that aims to address this potential problem.
- What does it mean to say that humans have “values”? Ayanna Howard says that training AI with human values may be possible. What problem does she anticipate, even if this goal is achieved? What does this tell us about the importance of diversity in science?