Computer science professor and AI expert Yejin Choi discusses training language models, the challenges of robots picking tools, and the role of universities in AI research.
Training a large language model is a challenging task in artificial intelligence research.
The future of AI research depends on the active involvement of universities and the exploration of alternative, compute-efficient methods.
Deep dives
AI's Opaque Nature and Need for Understanding
The podcast episode explores the opaqueness of current artificial intelligence (AI) systems, which lack transparency regarding how knowledge is encoded. The guest, Dr. E. J. Genshoy, highlights the challenge of understanding why AI models perform well in some tasks while making surprising mistakes in others. This lack of understanding on both artificial and human intelligence presents new intellectual problems. Prompt engineering, where slight changes in input yield different results, is discussed as a way to improve AI performance. However, reactions to prompt engineering results vary, with some emphasizing success and others focusing more on failure cases.
Scaling AI Models and Uncertainties
The podcast episode delves into the dramatic improvement seen in AI models like GPT-3 and the anticipation for future models like GPT-4. The guest reflects on initial skepticism about the usefulness of these models and highlights the uncertainty regarding their potential for further improvement. The conversation explores the challenges of evaluation, the role of larger scale models, and the need to explore alternative methods that are more compute-efficient and have less environmental impact. The podcast also touches upon specific use cases like math tutoring, where smaller specialized models can potentially surpass larger models in performance.
Concerns about Overreliance, Bias, and Misuse of AI
The episode raises concerns about the overemphasis on scale and prompt engineering, potentially leading to a lack of diversification in AI research. The conversation highlights the concentration of AI capabilities in a few tech companies, the need for greater openness and accessibility, and the importance of understanding the limitations and capabilities of AI systems. The potential misuse of AI, cognitive biases, and the impact on humanity are also discussed, stressing the need for ethical considerations and policy development to ensure AI benefits humanity without disastrous consequences.
Few people are better at explaining the science of artificial intelligence than Yejin Choi. She’s a computer science professor at the University of Washington, senior resource manager at the Allen Institute for AI, and the recipient of a MacArthur Fellowship. I thought her recent TED talk was terrific, and I was thrilled to talk to her about how you train a large language model, why it’s so hard for robots to pick tools out of a box, and why universities must play a key role in the future of AI research.