In this QuantaMagazine podcast, the host interviews Yejin Choi, a computer science professor at the University of Washington. They explore interesting topics such as whether AI must acquire embodiment and emotions to develop human-like common sense.
Currently, GPT-4 has shown some impressive "human-like awareness". In this podcast, Yejin Choi and host Steven Strogatz discuss the capabilities and limitations of chatbots and the large language models (LLMs) that build them. They explore whether AI can truly understand the world and the questions they answer.
As early as the 1960s, computer scientists dreamed of brain-inspired computers exhibiting human-like intelligence. With the rise of the internet, the emergence of large text datasets, and significant advances in computing power, we seem to have reached a crucial moment. Today's LLMs appear to possess something close to human intelligence.
Theories proposed by Sanjeev Arora of Princeton University and Anirudh Goyal, a research scientist at Google DeepMind, suggest that today's largest LLMs are not random parrots. As models grow larger and train on more data, their language abilities improve and combine skills in ways that suggest understanding, developing new capabilities.
Choi explains that LLMs simply read vast amounts of text and learn to predict the next word, but on a massive scale. They don't necessarily "reflect word-for-word" on the training data but can generalize to some extent. If text is repeated frequently enough in internet data, it will indeed memorize it verbatim.
The process of training LLMs can be boiled down to building a very large neural network with layers upon layers of neurons stacked together, then feeding internet data sequentially. The goal of the learning process is to predict the next word based on the sequence of previous words.
Although simple, this training method can produce powerful results, allowing LLMs to answer various questions in text. However, the process of training LLMs is vastly different from how humans understand the world. Humans learn through courses and curiosity, making assumptions about the world.
Choi believes that LLMs are a kind of "thought-emotion soup". They mimic human emotions and intentions because people do indeed invest emotions and intentions in their writing. But ultimately, they don't truly possess the genuine emotions that humans have.