
Автор: Erik Cambria
Издательство: Springer
Год: 2025
Страниц: 514
Язык: английский
Формат: pdf (true)
Размер: 18.1 MB
About half a century ago, AI pioneers like Marvin Minsky embarked on the ambitious project of emulating how the human mind encodes and decodes meaning. While today we have a better understanding of the brain thanks to neuroscience, we are still far from unlocking the secrets of the mind, especially when it comes to language, the prime example of human intelligence. “Understanding natural language understanding”, i.e., understanding how the mind encodes and decodes meaning through language, is a significant milestone in our journey towards creating machines that genuinely comprehend human language. Large language models (LLMs) such as GPT-4 have astounded us with their ability to generate coherent, contextually relevant text, seemingly bridging the gap between human and machine communication. Yet, despite their impressive capabilities, these models operate on statistical patterns rather than true comprehension. This textbook explores the current state of LLMs, their capabilities and limitations, and contrasts them with the aspirational goals of NLU. The author delves into the technical foundations required for achieving true NLU, including advanced knowledge representation, hybrid AI systems, and neurosymbolic integration. Containing exercises, a final assignment and a comprehensive quiz, the textbook is meant as a reference for courses on information retrieval, AI, NLP, data analytics, data mining and more.