Status: at work / Type of Theses: Master theses / Location: Dresden
Large Language Models (LLMs) have revolutionized natural language processing tasks, including automated question answering, summarization, and content generation. One critical aspect for ensuring the trustworthiness of these models is citation generation—the ability of LLMs to provide accurate and relevant references to support their outputs. However, this ability remains a challenge, particularly due to the lack of direct access to up-to-date and verified knowledge sources. Citation generation requires not only the generation of fluent text but also a strong alignment between the content produced and verifiable external sources.
This thesis focuses on the implementation and evaluation of different prompting strategies for citation generation with LLMs, comparing techniques such as zero-shot prompting, few-shot prompting, and chain-of-verification prompting. The study will further explore the impact of Retrieval-Augmented Generation (RAG), where external documents (e.g. scientific papers) are incorporated to assist LLMs in generating more precise and verifiable citations.
[1] Gao, T., Yen, H., Yu, J., Chen, D.: Enabling large language models to generate text with citations. In: Proceedings of the EMNLP 2023. pp. 6465–6488. (2023), https://aclanthology.org/2023.emnlp-main.398
[2] Dhuliawala, S., Komeili, M., Xu, J., Raileanu, R., Li, X., Celikyilmaz, A., Weston, J.: Chain-of-verification reduces hallucination in large language models. In: Findings of the ACL 2024. pp. 3563–3578. (2024), https://aclanthology.org/2024.findings-acl.212
[3] Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., Küttler, H., Lewis, M., Yih, W.t., Rocktäschel, T., Riedel, S., Kiela, D.: Retrieval-augmented generation for knowledge-intensive nlp tasks. In: Advances in NeurIPS. vol. 33, pp. 9459–9474. (2020), https://proceedings.neurips.cc/paper_files/paper/2020/file/6b493230205f780e1bc26945df7481e5-Paper.pdf