JavaScript is required to use this site. Please enable JavaScript in your browser settings.

Contact

AI Courses in Winter Semester 2024/2025

The following AI courses are provided by Prof. Dr.-Ing. Michael Färber in Winter Semester 2024/2025 at TUD Dresden University of Technology.

Course TitleModulesSWS
1Behind the Secrets of Large Language Modelstbd.2/2/0
2Erweitertes Komplexpraktikum Vertrauenswürdige Künstliche IntelligenzNF-B-510, INF-B-520, INF-MA-PR0/0/8
3Forschungsprojekt Vertrauenswürdige Künstliche IntelligenzINF-PM-FPA, INF-PM-FPG0/0/8
4Forschungsprojekt Vertrauenswürdige Künstliche Intelligenz (CMS)CMS-PRO0/0/12
5Großer Beleg Skalierbare Software-Architekturen für Data Analytics0/0/0
6Komplexpraktikum Vertrauenswürdige Künstliche IntelligenzINF-B-510, INF-B-520, INF-MA-PR0/0/4
7Teamprojekt Vertrauenswürdige Künstliche Intelligenz (CMS)CMS-LM-AI0/0/8


Lecture “Behind the Secrets of Large Language Models” (SecretLLM)

This course provides a practical and in-depth understanding of large language models that power modern natural language processing systems. Students will explore the architecture, training methodologies, capabilities, and ethical implications of LLMs. The course combines theoretical knowledge with hands-on experience to equip students with the skills necessary to develop, analyze, and apply LLMs in various contexts.

By the end of this course, students will be able to:

  1. Understand the architecture and key components of large language models.
  2. Analyze the training processes, including data collection, model optimization, and fine-tuning.
  3. Evaluate the performance and limitations of LLMs in different NLP tasks.
  4. Apply LLMs to real-world problems, such as text generation, summarization, and translation.
  5. Discuss the ethical considerations and societal impacts of deploying LLMs.

Prerequisites:

  • Introduction to Machine Learning or equivalent
  • Basic understanding of neural networks
  • Programming experience in Python

Potential topics for Komplexpraktikum etc.

Topic 0: Developing LLM-based Methods for Text Simplification

The goal of this project is to leverage advanced large language models, such as those similar to ChatGPT, to translate website content from ScaDS.AI and other websites into plain or simplified language (see project Klartext). By providing this in addition to the existing German and English versions, the project aims to break down language barriers and enhance public engagement. This project significantly contributes to inclusion by enabling a broader audience, including non-experts and those with varying language proficiencies, to participate in scientific and cultural conversations.

What are the tasks?

  • Designing and implementing LLM-based methods for text simplification.
  • Evaluating these methods for use cases such as the translation of text into Easy/Plain Language.

What prerequisites do you need?

  • Strong interest in research on natural language processing, especially LLMs.
  • Good programming skills.

Topic 1: Designing and Executing a Large-Scale User Study on Scientific Text Simplification

This topic is about making scientific texts more understandable. The goal is to automatically rewrite or translate academic articles so they become clear not only to field experts but also to researchers from other disciplines and interested laypeople. This role involves planning and conducting a user study. It’s a unique chance to actively engage in a project that could transform how we interact with scientific knowledge. You’ll gain experience in research methodology and user study design, directly contributing to making science more accessible to a broader audience.

What are the tasks?

  • Designing a comprehensive user study (e.g., select and preprocess the scientific texts for the participants).
  • Collecting and analyzing the response from the participants (e.g., who summarized or simplified scientific texts in their own words) so that the responses can be used as a ground truth (“perfect texts”) for AI models.
  • If you have experience with programming: Implement generative AI models (e.g., GPT-based) which can summarize or simplify scientific texts automatically.

What prerequisites do you need?

  • A passion for making scientific knowledge accessible to a broader audience.
  • Strong interest in research, user study design, and data analysis.
  • Good organizational skills to effectively design and manage a large user study (funded by the chair).
  • Basic programming skills.

Topic 2: Stock Market Predictions through Deep Learning

This topic focuses on a collaboration with Orca Capital, a company specializing in financial markets. Together with Orca Capital, a Munich-based startup has developed a runing system that predicts the stock prices of certain companies based on a continuous stream of news, such as rising/falling prices and volatility. This system utilizes deep learning and natural language processing methods, including pretrained language models. The students will work on further developing and enhancing the system, using real-world financial data and industry contacts. Possible enhancements include applying the latest language models (LLMs) and techniques to make the predictions more explainable (explainable AI).

What are the tasks?

  • Developing extensions and improvements of the system, using the latest findings in deep learning and natural language processing.
  • Evaluating the system’s performance and making its predictions more interpretable, integrating methods from the field of explainable AI.

What prerequisites do you need?

  • Good programming skills in Python.

Topic 3: Large Language Model-enhanced Graph Message Passing Network for Link Prediction

This topic is about advancing AI-based recommendation methods through the integration of large language models and graph message passing networks. The project aims to revolutionize how we predict and understand linkages within academic citation networks.

What are the tasks?

  • Implementing and testing algorithms for link prediction, community detection, node classification, and potentially other graph-supervised learning tasks.
  • Exploring the trade-offs between the utilization of textual and structural features in link prediction algorithms, and devising methods to efficiently combine these features.

What prerequisites do you need?

  • A strong interest in machine learning, natural language processing, or graph theory.
  • Proficiency in programming, preferably in Python, with experience in PyTorch or TensorFlow.
  • Eagerness to engage with state-of-the-art research in link prediction and text mining.

Topic 4: Extending the RDF Knowledge Graph SemOpenAlex.org

This topic is about working on SemOpenAlex, a comprehensive RDF knowledge graph that includes over 26 billion triples related to scientific publications, authors, institutions, journals, and more. This open-access initiative offers data through RDF dump files, a SPARQL endpoint, and the Linked Open Data cloud, enhancing the visibility and accessibility of scientific research.

What are the tasks?

  • Keeping SemOpenAlex up-to-date by updating its schema according to changes in the OpenAlex database and performing periodic updates to the RDF database.
  • Expanding SemOpenAlex, e.g., by introducing author name disambiguation, integrating representations of code repositories like GitHub, and linking to other databases and knowledge graphs such as LinkedPaperWithCode.com, Wikidata, and DBLP.

What prerequisites do you need?

  • Basic understanding of RDF and enthusiasm for semantic web and open data.
  • Programming skills in Python, which are critical for various tasks including database maintenance and development.

Topic 5: Fusing RDF Knowledge Graphs with Deep Learning for Advanced Recommender Systems

This project seeks to expand AutoRDF2GML, an open-source framework acclaimed for converting RDF data into specialized representations ideal for cutting-edge graph machine learning (GML) tasks, including graph neural networks (GNNs). With its automatic extraction of both content-based and topology-based features from RDF knowledge graphs, AutoRDF2GML simplifies the process for those new to RDF and SPARQL, making semantic web data more accessible and usable in real-world applications.

What are the tasks?

  • Adapt AutoRDF2GML to process a broader range of RDF knowledge graphs, allow a flexible integration of data sources from the Linked Open Data cloud.
  • Redesign the AutoRDF2GML interface to be more intuitive and user-friendly, enabling a seamless experience for both new and experienced users.
  • Boost the framework’s automation capabilities to simplify the setup and execution processes, making it easier to generate and use graph machine learning datasets efficiently.

What prerequisites do you need?

  • Proficiency in Python, with a foundational understanding of RDF, SPARQL, and graph machine learning concepts.
  • An enthusiastic interest in the intersection of semantic web technologies and deep learning.
funded by:
Gefördert vom Bundesministerium für Bildung und Forschung.
Gefördert vom Freistaat Sachsen.