Knowledge-enhanced natural language processing

Frisoni, Giacomo (2024) Knowledge-enhanced natural language processing, [Dissertation thesis], Alma Mater Studiorum Università di Bologna. Dottorato di ricerca in Computer science and engineering, 36 Ciclo.
Documenti full-text disponibili:
[img] Documento PDF (English) - Accesso riservato fino a 15 Maggio 2025 - Richiede un lettore di PDF come Xpdf o Adobe Acrobat Reader
Disponibile con Licenza: Creative Commons Attribution Non-commercial No Derivatives 4.0 (CC BY-NC-ND 4.0) .
Download (60MB) | Contatta l'autore

Abstract

Language is a fundamental part of being human. It allows people to communicate thoughts, express ideas, create memories, and build mutual comprehension. Developing machines capable of understanding language automatically has been a long-standing dream in artificial intelligence. Over the past decade, dramatic and unprecedented breakthroughs in natural language processing have been driven by transformer-based networks scaled on progressively larger parameter counts and training data. Large language models have recently become prominent, revealing early signs of general intelligence and ushering in a new era of human-machine interaction. However, although pre-trained language models could effectively store numerous facts encountered at training time, their knowledge awareness is still far from satisfactory. Current models grapple with lexical superficiality, human reporting bias, hallucination, poor reasoning, black-box behavior, and computational complexity. These open issues have led researchers to question the adequacy of scaling as the sole condition for scientific progress. A growing number of influential thinkers recognize that the key direction for substantial advancements lies in the incorporation of semantic-level abstractions and pertinent context derived from external knowledge sources. In this dissertation, we study knowledge-enhanced natural language processing, presenting multiple techniques for knowledge extraction, representation, and injection. We show that semantic parsing, graph representation learning, retrieval augmented generation, multi-modal signals, and back-propagation through combinatorial solvers are integral components of the solution. Overall, this work represents a significant step toward delineating a new generation of language models that prioritize meaning, reasoning, and interpretability.

Abstract
Tipologia del documento
Tesi di dottorato
Autore
Frisoni, Giacomo
Supervisore
Co-supervisore
Dottorato di ricerca
Ciclo
36
Coordinatore
Settore disciplinare
Settore concorsuale
Parole chiave
natural language processing; large language models; knowledge extraction; semantic parsing; knowledge representation learning; graph neural networks; knowledge injection; retrieval-enhanced language models
URN:NBN
Data di discussione
24 Giugno 2024
URI

Altri metadati

Gestione del documento: Visualizza la tesi

^