
"Discussion with Nate Soares on a key alignment difficulty" by Holden Karnofsky
LessWrong (Curated & Popular)
00:00
The Importance of Reflection in AI Research
Natus: In order to do certain kinds of research effectively, you need to keep asking questions like what am I actually trying to do here and why? What is my big picture goal? Which are questions that might change your aims in some important sense? It's relatively hard to pick up a robust pewter avoidance pattern even without actively screwed up data points. The less granularly you're making your AI imitate a human, the more you're doing something like, outcomes-based trial and error where the AI could be achieving human-like end products with very alien intermediate steps.
Play episode from 11:18
Transcript


