Project description:
Large
Language Models (LLMs) are well-known for knowledge acquisition from
large-scale corpus and for achieving SOTA performance on many NLP tasks.
However, they can suffer from various issues, such as hallucinations, false
references, made-up facts. On the other hand, Knowledge Graphs (KGs) can store
enormous amounts of facts in a structured and explicit manner. However, unlike
LLMs, formulating KGs is a laborious process, and querying KGs might be
computationally demanding. An interesting research question is then the
following: How to combine KGs and LLMs such that LLMs provide answers based on
facts and do not hallucinate in any way. We will use a well-established KG
(e.g., from biology or chemistry) and focus entirely on the combination of LLMs
and KGs.
In this
thesis: (a) you will study the techniques for combining LLMs and KGs, (b) you
will formulate and code your own LLM+KG, (c) you will design and carry out
evaluations for your LLM+KG.
Literature (examples):
Prerequisites: