Open topics for theses and practical courses
Unless otherwise specified, all topics are available as practical course (P1/P2), bachelor or master thesis.
Work group:
|
Type:
|
A Unified Perception of Density for Clustering [Master Thesis]
In this master thesis, we explore the hierarchical properties of results from density-based clustering with respect to the discrete minPts parameter instead of the frequently regarded epsilon. For that, we incorporate recent research [*] and insights from the field of persistent homology.
[*] Beer, Anna, et al. "Connecting the Dots--Density-Connectivity Distance unifies DBSCAN, k-Center and Spectral Clustering." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023. DOI
Contact: Anna Beer
Internal Evaluation for Density-based Clustering
We develop a new internal evaluation measure for clusterings based on recent research [*] and compose a survey of current evaluation methods. All methods will be compared systematically and extensively for different settings and concepts of density. Furthermore, we will investigate state-of-the-art benchmark data sets w.r.t. their suitability for the evaluation of density-based clustering algorithms considering our new measure.
[*] Beer, Anna, et al. "Connecting the Dots--Density-Connectivity Distance unifies DBSCAN, k-Center and Spectral Clustering." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023. DOI
Contact: Anna Beer
Clustering on Trajectories: Finding Dynamic Domains in Molecular Dynamics Data [Practical Course or Master Thesis]
The data regarded in this thesis is from the field of Molecular Dynamics (MD) and describes the positions of a protein's atoms over time. Clustering allows to detect dynamic domains, i.e., rigid structures in the protein, which is important for understanding the protein folding process. In this work, an existing but yet unpublished clustering idea for MD data should be extended, improved, and tested extensively on MD as well as other data.
Contact: Anna Beer
3d-Shape Descriptors and Molecular Descriptors for Clustering [Practical Course or Master Thesis]
The data regarded in this thesis is from the field of Molecular Dynamics (MD) and describes the positions of a protein's atoms over time. Clustering timesteps allows to detect different states of a protein, which is important for understanding the protein folding process. E.g., these states can serve for a Markov State Model of the protein's trajectories.
In this work, we compare results based on the usually applied molecular descriptors with 3d-Shape descriptors that were recently developed for general point clouds.
Contact: Anna Beer
Permutation-aware Graph Similarities
Molecular structures are often represented by graphs and comparing them is a common task. Graph models are usually designed to be permutation invariant. The structure of molecules, however, imposes certain restrictions on the arrangement of neighbors. For example, so-called stereoisomers have the same molecular graph but differ in shape.
The Weisfeiler-Leman algorithm is an iterative vertex coloring scheme, originally developed to distinguish non-isomorphic graphs and often used in graph kernels. It also provides the theoretical foundation for the message passing in many graph neural networks. It iteratively computes node colors by using the color of the previous iteration and the multiset of colors of the neighboring nodes. Using a multiset, the order of the elements typically does not matter.
In molecular graphs, however, a different ordering of neighbors sometimes leads to functional differences. To consider this, an approach that is only invariant to certain permutations would be more suitable. The goal of this project is to extend the Weisfeiler-Leman algorithm to respect such differences.
Requirements: Python or Java, interest in graph theory
Contact: Nils Kriege, Franka Bause
Graph Neural Networks and Molecular Fingerprints
Early graph neural networks (GNNs) were motivated by circular chemical fingerprints generalizing their concept by introducing learnable parameters. However, recent studies show that current state-of-the-art GNN architectures are still outperformed in molecular property prediction by concrete implementations of circular fingerprints combined with a multilayer perceptron. The task of this project is to explore the reasons for this through a detailed analysis of both methods. You will investigate a concrete implementation of a chemical fingerprint and develop a new neural version starting by reproducing the fingerprint and adding neural modules stepwise. Different stages should be investigated in an ablation study.
Contact: Nils Kriege
Permutation Invariant Neural Architectures
Machine learning techniques benefit from taking the specific properties of the considered examples into account. In various domains, we have to deal with sets of elements. In a set, the order of elements does not matter. Consequently, the order of elements in a set given as input to a (well-designed) machine-learning algorithm for sets should not affect its output, a property referred to as permutation invariance. Neural architectures satisfying this property are well investigated. In this project, we would like to develop extensions achieving invariance regarding specific families of permutations.
Contact: Nils Kriege, Sebastian Tschiatschek
Explaining Graph Neural Networks
Graph neural networks (GNNs) have been widely used in different applications like recommender systems and drug discovery. It is of major interest to interpret the results and explain deep learning methods to people working in non-technical domains. The explainability of machine learning methods has been well-investigated in the literature. In this project, we focus on methods that can be extended or applied to graph-structured data and GNNs. The goal is to (1) summarize and replicate state of-the-art explainability methods, (2) investigate their limitations and applicability to graphs, and (3) design a new method/framework for explaining GNNs with a special focus on drug discovery or recommender system applications.
Contact: Ghaith Mqawass, Nils Kriege
Expressivity of Graph Invariants
A function is graph invariant if it leads to the same output for any two graphs that are isomorphic. The Weisfeiler-Leman algorithm provides such a graph invariant function, but cannot distinguish some simple non-isomorphic graphs (for example, a hexagon from two disconnected triangles). The goal of this project is to investigate different representations of node neighborhoods and relate them according to their expressivity. We would like to compare graph invariants experimentally on benchmark datasets and, if possible, also find theoretical connections between them.
Requirements: Python or Java, interest in graph theory
Contact: Franka Bause, Nils Kriege
Gradual Weisfeiler-Leman Refinement
The Weisfeiler-Leman algorithm is an iterative vertex coloring scheme, originally developed to distinguish non-isomorphic graphs and often used in graph kernels. The vertices of a graph are iteratively colored based on their previous colors and the multiset of colors of their neighboring vertices. While this leads to an algorithm that can distinguish many non-isomorphic graphs, the similarity function on the vertices implied by the resulting color hierarchy is not fine-grained enough, when used in a graph kernel. In the gradual variant of the Weisfeiler-Leman algorithm (GWL), this problem is tackled by defining the set of recoloring functions that, while not necessarily injective, lead to the same stable coloring as the original WL algorithm. The benefits of this method were shown exemplary, using k-means to define the recoloring function. The goal of this project is to re-implement the existing GWL-framework in Python and extend it using other possible functions that also fulfill the specific criteria needed. The extension to graphs with arbitrary vertex attributes and the usage of multiple color hierarchies are further directions that can be explored.
Requirements: Python, interest in graph learning
Contact: Franka Bause, Nils Kriege
Building self-explanatory transparent models
Explainable Artificial Intelligence (XAI) focuses on providing explanations for domain experts and data scientist. However, 'lay people', i.e., people without or only with limited technical knowledge, are often subject to the consequences of deploying algorithmic decision-making systems in public institutions, such as the COMPAS recidivism model. Finding new ways to render machine learning models comprehensible for lay people is therefore a current and highly relevant challenge. This project will focus on approaches for building machine learning models that are in themselves interpretable and explainable for lay people, soliciting slight or no external explanations. It will involve both programming and the conducting of user studies with lay people.
Contact: Timothée Schmude, Sebastian Tschiatschek
Causal Structure Learning for Questionnaires
Many research question are not only statistical but rather causal in nature. Two variables may be statistical dependent on each other, but do not not cause each other. For example, the relative number of Nobel laureates for a country strongly correlates with its annual chocolate consumption. However, increasing the national chocolate consumption, does obviously not produce more outstanding scientists. It is a key challenge for machine learning to move from learning purely statistical relations to causal ones. In the absence of randomized trials, causal structure learning focuses on identifying which variables in a given dataset are directly causally related and how these direct causal relations form a structure over which also indirect causal relations exist. These dependencies can be encoded by a Directed Acyclic Graph (DAG) and the first step in performing causal inference is to determine such underlying causal graph.
In this project you will make yourself familiar with causal structure learning in machine learning. You will benchmark recent algorithms to find underlying causal relations in datasets and apply them to questionnaire data from the social and political sciences.
Students working on this project need basic background knowledge in machine learning or probability, good programming skills in Python, and the desire to implement and test causal discovery algorithms.
Contact: Simon Rittel, Sebastian Tschiatschek
Attentive pixel prediction (Reinforcement Learning)
Auxiliary prediction tasks have been the source of many recent performance improvements for reinforcement learning agents, particularly when working with image inputs. These tasks are usually provided as expert knowledge by the algorithm designer and force the agent to pay attention to various aspects of the environment which might be helpful to solving the underlying task. In this project you will investigate whether such prediction tasks can also be learned by the agent itself. For this, you will first visualize what an RL agent actually "attends to" through investigating heatmaps of their first network layer. Then you will devise a method allowing the agent itself to decide which pixels of the state it wants to learn to reconstruct.
Required skills: Good programming skills, basic background in deep learning
Contact: Timo Klein, Sebastian Tschiatschek
Machine learning based climate model weighting
Climate models, global and regional, are based on numerical and physical equations and parametrizations and, thus, are computationally expensive especially when trying to simulate more than 10 years. Based on the interpretation of equations (e.g. topography following atmosphere) and the parametrizations for e.g. clouds results of climate models vary in their skills. This is especially the case in regions with complex topography.
In the past years the idea of mixing different climate model simulations based on their respective skills in e.g. reproducing history precipitation has emerged. Thus, generating a sort of mean climate simulation consisting of information of different climate models. This is called weighted model mixing (e.g., Brunner et al., 2020; Rana et al. 2013). The best performing model with respect to a specific observation type thus has the most influence on the resulting simulation.
With the emerging skill of different machine learning models the statistical weighting (Rana et al., 2013, Brunner et al., 2020, Merrifield et al., 2020, Hamill and Scheuerer (2018)) often has lower skills compared to e.g. a weighted Convolutional Neural Network (Bertrand et al., 2021). Other approaches can have shown good skills, too (Wong et al. 2021, Sun and Archibald (2021), Segupta et al. (2021), Amos et al. (2021).
The idea of the proposed thesis is to implement two methods, a classical statistical as baseline and a machine learning model, to combine selected GCM and RCMs with the weighted algorithm based on their respective skills.
Contact: Irene Schicker (ZAMG), Katerina Schindlerova
The Complexity of Computing the Graph Edit Distance
The graph edit distance (GED) quantifies the dissimilarity of two graphs as the minimum cost of a sequence of edit operations turning one graph into the other. The problem complexity depends on the choice of the edit operation costs and the properties of the considered graphs. It has been shown that the classical NP-hard subgraph isomorphism problem (SI) and the maximum common subgraph problem (MCS) can be reduced to computing the GED. As a consequence, computing the GED also is NP-hard. However, for SI and MCS a fine distinction of the complexity depending on parameters such as treewidth, degree, and their combination is known. The goal of this project is to investigate the complexity of computing the graph edit distance in restricted graph classes identifying polynomial-time solvable and NP-hard cases.
Requirements: Strong interest in graph algorithms and theoretical computer science
Contact: Nils Kriege
Reinforcement Learning for improving mental health treatments
In this project/thesis you will study the potential of using reinforcement learning (RL) for improving mental health treatments. In particular, you will investigate the potential of RL algorithms for optimizing exposure therapy protocols, which are usually used for the treatment of anxiety disorders, such as phobias.
Specifically, you will conduct simulation studies in which a computational associative learning model (such as the latent cause model) will be a stand-in for the patients, and where an RL algorithms will make modifications to the exposure protocol with the goal of robustly extinguishing negative associations, which play a crucial role in anxiety disorders. The investigation can start with standard model-free RL approaches, but is then planned to proceed with model-based augmentations, such as equipping the RL algorithm with "Theory-of-Mind"-type capabilities.
Students working on this project need basic background knowledge in machine learning, solid programming skills, a desire to work on reinforcement learning, and willingness to learn about models of human cognition.
Contact: Sebastian Tschiatschek
Information-theoretic measures and the EEG time series of depression patients [Master Thesis]
Based on the EEG recordings database of depression patients, the goal is to utilize information-theoretic measures to assess the distances among the EEG signal of the patients under different treatment conditions with the intention to apply the results to a related classification problem. The interested student will be provided more details.
The work will be done within the international research project "Learning Synchronization Patterns in Multivariate Neural Signals for Prediction of Response to Antidepressants", which is a joint research project of University of Vienna and of the Czech Academy of Sciences
Contact: Katerina Schindlerova, Claudia Plant
Efficient algorithms for uncertain graphs [Master Thesis]
Uncertain graphs represent an important data model for real-world networks, where one can assign an independent probability of existence on every edge. Uncertainty, represented by the edge probability, may arise due to several reasons, e.g., errors in measurements, data integration from inconsistent and ambiguous sources, lack of precise information needs, inference and prediction models, and explicit manipulation for privacy purposes. Due to their rich expressiveness and utility in a wide range of applications, uncertain graphs have prompted a great deal of research by the data mining research community.
This work will mainly focus on designing efficient algorithms for selecting and modifying the probability of a given number of edges in an uncertain graphs. The developed algorithm will be evaluated against other existing methods. An experimental study will be conducted on real-world datasets. Students working on this project need basic knowledge in graph theory and algorithms, and solid programming skills in Python, C/C++ or Java.
Contact: Yllka Velaj, Claudia Plant
Common Subgraph Problems in Tree-Like Graphs
For two given graphs, G and H, the maximum common subgraph problem (MCS) asks for the largest graph contained in both G and H. An important application occurs in cheminformatics, where the similarity of molecular graphs needs to be quantified. Unfortunately, MCS is NP-hard unless additional constraints regarding the properties of G, H, and the subgraph are enforced. A polynomial-time solvable case requires that all these graphs are trees. Known negative complexity results for classes of tree-like graphs set narrow boundaries for generalization.
The project aims to investigate particular tree-like graph classes to design, implement, and experimentally evaluate new polynomial-time algorithms. New NP-hardness proofs for specific cases should refine the complexity status further.
Requirements: Strong interest in algorithms and graph theory
Contact: Nils Kriege
Dynamic Orbit Partition [Practical Course or Master Thesis]
A graph isomorphism is a mapping between the nodes of two graphs that preserves the edges. An automorphism is an isomorphism of a graph to itself. The (non-trivial) automorphisms intuitively represent the symmetries of a graph. Two vertices are considered equivalent if there is an automorphism that maps one vertex to the other. The corresponding equivalence classes are referred to as orbit partition. The orbit partition is of interest for symmetry breaking techniques for various combinatorial problems. Identifying and pruning symmetric branches in the search tree based on the symmetries of the input instance can lead to drastic speed-ups. This project aims to develop dynamic algorithms for maintaining the orbit partition when the input graph changes. The focus of the project can be on the development and theoretical analysis of algorithms or their implementation and experimental evaluation.
Requirements: Strong interest in algorithms
Contact: Nils Kriege, Christian Schulz, Kathrin Hanauer
Interpretable and Explainable Deep Learning
In this project you will study novel approaches for learning interpretable and explainable deep learning models with the goal of supporting users with different skill levels and knowledge. The models will be evaluated on applications with large societal impact, e.g., labour market data.
Contact: Sebastian Tschiatschek
Dynamic Information Acquisition in Questionnaires
In many areas of our life questionnaires/a collection of measurements are used to gather information for decision making, e.g., questionnaires are used to measure the success of marketing campaigns or customer satisfaction, or medical tests are conducted to decide on the best treatment for a patient. Common to those applications is that many questions need to be answered or many tests must be conducted. Often this is not because all the information is needed for deriving a decision but rather because the course of action is standardized (e.g., a survey always consists of the same questions).
In this project you will evaluate and develop methods for adaptive information acquisition that acquire the information necessary for decision making sequentially and adaptively, e.g., not all questions in a questionnaire are asked in all cases and the order of the questions can be changed. You will consider and evaluate recent machine learning models for this problem but also investigate how these methods can be extended to account for dependencies in the information that can be required, e.g., dependencies that only allow for a specific order of some questions of a questionnaire, the influence of how questions are asked on the answers, etc.
Contact: Sebastian Tschiatschek
Deep Probabilistic Clustering for Heterogeneous and Incomplete Data
Clustering is the task of finding groups in a data set. As an unsupervised learning approach it is important for exploratory data analysis, e.g., for finding commonalities between patients suffering from some disease. Often such data is high-dimensional and contains complicated non-linear relationships, which makes it difficult to apply standard clustering approaches. Deep clustering algorithms have been proposed recently to approach this kind of setting by combining clustering and deep learning in a common framework. An open problem that has not been sufficiently addressed in recent deep clustering research is incomplete and heterogeneous (e.g. survey data, blood tests, MRI images, etc.) data. Incomplete data is often an issue in practice, especially for health data where not everything about a patient’s history is known.
In this project you will study the clustering of incomplete data with heterogeneous data types with a particular focus on missingness and on determining which information is relevant for cluster assignments. You will develop an algorithm based on previous work and evaluate it on synthetic and real-world data. Some research questions of interest would be for instance: How do cluster assignments change when given new information? Or, given an uncertain cluster assignment (e.g. sick or healthy patient) which information would we need to query to make a decision?
Contact: Sebastian Tschiatschek
Multi-agent Reinforcement Learning under Mismatch
Enabling intelligent agents to collaborate effectively despite mismatch in perception or capabilities.
Contact: Sebastian Tschiatschek
Multi-agent Teaching Primitives
Enabling multiple interacting agents to convey their knowledge and skills through teaching.
Contact: Sebastian Tschiatschek
Abstraction in Reinforcement Learning
Learning abstractions of large state and action spaces in order to facilitate more efficient planning and exploration.
Contact: Sebastian Tschiatschek
Bivariate causal measures: Experiments with existing causal methods in python on synthetic and real data [Practical Course]
Bivariate causal measures: Experiments with existing causal methods in python on synthetic and real data.
Contact: Katerina Schindlerova
Oracle analysis of distant supervision errors
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Weakly supervised discourse relation prediction
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Incomplete Schema Relation Clustering
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Weakly supervised learning with latent class predictions
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Gradient matching for semi-supervised learning
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Threshold-finding for knowledge-base completion using Gaussian processes
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Path-based knowledge-base completion
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Better sentence representations based on BERT
Please see [Moodle] for a detailed description of this and other Natural Language Processing topics.
Contact: Benjamin Roth
Explainable Policies for Game Play [Master Thesis]
Reinforcement learning has shown remarkable successes in playing board games like Go but also in playing computer games like DOTA. While these results are impressive, the learned strategies (policies) are typically not directly interpretable by humans. However, interpretability is important for instance to enable better collaboration of humans with AI agents or to assess safety of the learned behavior.
In this project you will make such learned policies easier to understand by humans by incorporating constraints on their structure. You will develop and implement a novel policy training algorithm and evaluate it on simple computer games.
Students wanting to work on this topic are expected to have solid coding skills in Python, basic knowledge in PyTorch or TensorFlow, and be curious to learn about reinforcement learning and its applications.
Contact: Sebastian Tschiatschek
Learning Composable Policies [Master Thesis]
Reinforcement learning has shown remarkable successes in application domains like game play and robotics but suffers from large sample complexity, limiting its broader usage. A promising approach for reducing the sample complexity is to learn skills that can be reused when solving a series of related tasks. In the context of robotics, such skills could for instance correspond to lifting a leg or grepping an object.
In this project you will develop and implement a learning algorithm algorithm for such composable policies based on recently proposed Successor Options. You will evaluate this algorithm on several benchmark tasks and compare it to other state of the art algorithms.
Students wanting to work on this topic are expected to have solid coding skills in Python, basic knowledge in PyTorch or TensorFlow, and be curious to learn about reinforcement learning and its applications.
Contact: Sebastian Tschiatschek
The Cost of Feedback
Reinforcement learning has shown remarkable successes in application domains like game play and robotics. These successes where achieved by performing large numbers of interactions of a learning agent with the environment. In these interactions, the agent constantly receives information about rewarding behavior. However, in many applications involving human users (e.g., personal assistants), the reward information is explicitly or implicitly provided by these human users. Therefore, this information must be treated as a scarce and valuable resource, and the (cognitive) cost of providing this information should be accounted for.
In this project you will study the cost of providing different types of feedback used for reinforcement learning in user studies. To this end, you will implement a simple reinforcement learning environment for which a human user can provide reward information in various forms (e.g., comparisons, sorted lists, etc.) and compare the cost of providing this feedback as well as the usefulness of the feedback for training the reinforcement learning agent. Depending on whether you are doing a P1/P2/Bachelor Thesis/Master Thesis, the scope of the project will be adjusted.
Students wanting to work on this topic are expected to have solid coding skills in Python, basic knowledge about the development of interactive web pages, and be curious to learn about reinforcement learning and its applications.
Contact: Sebastian Tschiatschek
Reinforcement Learning for Optimizing Electronic Circuit Design [Practical Course]
Electronic circuits are parts of all electric devices we use on a daily basis. When developing a new electric device, tailored electronic circuits have to be constructed for the specific use case, an often time-consuming process. Therefore it is appealing to aim to automate this process. Reinforcement learning is a promising approach therefore and has shown remarkable successes in application domains like game play and robotics but is so far rarely used for designing and optimizing electronic circuits.
In this project you will build the basis for using reinforcement learning for electronic circuit design and optimization. In particular, you will wrap an existing electronic circuit simulator in a reinforcement learning environment such that it can be easily interacted with by standard reinforcement learning algorithms. You will then implement and evaluate standard reinforcement learning algorithms for designing and optimizing simple electronic circuits, e.g., current sources.
Students wanting to work on this topic are expected to have solid coding skills in Python and be curious to learn about reinforcement learning and apply it in a real-world application. Some background in electronic circuit design is advantageous but not a prerequisite.
Contact: Sebastian Tschiatschek
2D pharmacophore descriptors for machine learning and similarity searches
Pharmacophores are an abstract description of molecular features that are necessary for the molecular recognition of drugs by a biological macromolecule (usually proteins). The molecular features constituting a pharmacophore can be divided into different classes according to the nature of their non-bonding interaction with complementary parts of the macromolecular target receptor. Every molecule has a characteristic pharmacophoric feature pattern in terms of number, type and mutual distance of its features. Different molecules that are recognized by the same biological target usually show a high similarity regarding their feature pattern. A database search for molecules which have a pharmacophoric profile that is similar to a given query pharmacophore can thus reveal novel molecules that have a high chance to bind to the same receptor for which the query was created. To be able to search millions of molecules in acceptable time, a suitable in-memory pharmacophore representation is required which allows for a fast similarity assessment. In cheminformatics, an often employed representation of molecular structures which fulfills these requirements are fixed-size binary fingerprints (bit strings) where similarity calculations essentially boil down to fast logical bit-set operations.
The goal of this project is to implement such a binary encoding also for pharmacophores that can be used for fast pharmacophore similarity searches and for the development of machine learning models. The intended algorithm for converting an arbitrary pharmacophore into a fixed-size bit representation is exemplified in the following figure:

The implementation will be carried out on top of the open source cheminformatics toolkit CDPKit (https://github.com/aglanger/CDPKit) which already provides all required functionality for the I/O of molecular structures and pharmacophore generation.
Requirements: C++ (preferable) or Python, basic knowledge about the fundamentals of organic chemistry advantageous
Contact: Thomas Seidel, Nils Kriege
Causal inference for wind turbine extreme events [Bachelor or Master thesis]
This work will focus mainly on programme implementation of algorithms for causal detection among extreme event processes and their testing on real data. The causal inference will be tested among the wind turbine time series as well as other relevant meteorological time series, as provided by Austrian ZAMG (Die Zentralanstalt für Meteorologie und Geodynamik). More details to the interested student will be provided on request.
Contact: Katerina Schindlerova
Clustering of spatio-temporal climatological data [Master thesis]
This work will focus on the integration of massive amounts of data of heterogeneous spatial and temporal resolution of spatio-temporal climatological data provided by ZAMG. An effective information-theoretic objective function supporting different length of time series and different spatial resolutions will be applied. A joint low-dimensional vector space embedding of all data will be learned, which will be then used as a model for data compression by clustering. Measurement sites with similar characteristics are in this space close to each other and measurement sites with different characteristics will be far apart from each other in this vector space. Deep neural nets and greedy optimization techniques such as iterative least square methods will be considered to learn this space. A clustering algorithm will be developed, which works with this representation of the data and iteratively refines it during the clustering process.
Contact: Katerina Schindlerova, Claudia Plant
Reinforcement Learning from Implicit and Explicit Feedback
Recent advances in reinforcement learning have enabled super-human performance of AI agents in many applications, e.g., game play. In this project you will work on building theory and/or models for reinforcement learning from implicit and explicit feedback. In particular, you will consider the setting in which no rewards are observed during execution of the agent but only cumulative reward information is obtained at the end of the execution. As a practical example, assume playing a computer game in which you cannot observe the score while you play the game but only see your achieved score at the very end. This setting has important applications in human-in-the-loop settings in which feedback can be expensive to obtain.
Students working on this project need basic background knowledge in machine learning, solid programming skills in Python, and the desire to work on reinforcement learning.
Contact: Sebastian Tschiatschek
Machine Learning for Personalized Education [Practical Course or Bachelor Thesis]
In this project you will work on predicting students' responses to mathematical questions using machine learning models. This is important to assess a student's skills and provide personalized educational resources. We use real-world data provided by the "Diagnostic Questions: Predicting Student Responses and Measuring Question Quality" challenge at NeurIPS'20. The challenge consists of four different tasks from which a subset can be selected for the project, depending on background knowledge and programming skills.
Students working on this project need basic background knowledge in machine learning, programming skills in Python, and the desire to develop and test machine learning models.
Contact: Sebastian Tschiatschek
Probabilistic Models of Multimodal Distributions
Many real-world phenomena can only be accurately described by multimodal distributions, e.g., the hourly power consumption of a household (typically) is bimodal, with peaks in the morning and the evening. However, learning deep generative models of these multimodal distributions, e.g., using variational auto encoders, still remains challenging. In this project you will develop theory and/or models for multi-modal distributions based on variational autoencoders. You will evaluate these models on benchmark and real-world datasets.
Students working on this project need basic background knowledge in machine learning, programming skills in Python, and the desire to develop and test machine learning models.
Contact: Sebastian Tschiatschek
Reward Inference for Sequential Decision Making from Diverse and Implicit Feedback [Master Thesis]
Automated sequential decision making is an important application of machine learning systems in which such a system needs to select a sequence of actions step by step to optimize a reward/utility function. For instance, in autonomous driving, such a system needs to execute a sequence of steering, breaking and acceleration actions, or in a medical intensive care setting, such a system needs to execute a sequence of measurement and treatment actions.
One challenge in realizing such automated sequential decision making systems is the definition of the reward/utility function. For example, in autonomous driving it is hard to specify all the factors which define good driving behavior. In such settings, automatically inferring the reward/utility function from users’ feedback can be beneficial.
This project investigates approaches for reward/utility inference from diverse and implicit feedback, building on ideas for inverse reinforcement learning, active learning, implicit feedback, etc.
Interested students are expected to have solid mathematical and machine learning skills, and have experience in Python and deep learning (using PyTorch or TensorFlow).
Contact: Sebastian Tschiatschek
Imitation Learning Under Domain Mismatch
Reinforcement learning has been successfully used to solve certain challenging sequential decision making problems in recent years. The employed techniques commonly require (i) huge amounts of interactions with the envirnoment and (ii) clearly specified reward signals to work well. In many applications however, one or both of these requirements are not met. In such cases, imitation learning can be an efficient approach to sequential decision making problems: an expert demonstrates near-optimal behavior and a learning agent attempts to mimic this behavior.
This project considers imitation learning in settings in which there is some form of mismatch between the expert demonstrator and the learning agent. The scope of the project is to study how existing algorithms perform in this setting and proposes modifications to existing algorithms to achieve better performance.
Students wanting to work on this topic are expected to have a basic understanding of machine learning techniques, solid knowledge of Python and basic knowledge of deep learning libraries (PyTorch or TensorFlow).
Contact: Sebastian Tschiatschek
Selecting Sequences of Items for Non-monotone functions [Bachelor or Master Thesis]
Many applications involve the selection of sequences of items, e.g., recommender systems and personalization in MOOCs. Common to many of these applications is that the order of the selected items is not random but that items selected early, influence which items are selected later. The problem of selecting sequences of items has been studied in various settings, including that in which dependencies between items are modeled using graphs and monotone submodular functions.
This project aims at extending these settings to cover the case of non-monotone submodular functions, by proposing new algorithms and analyzing their properties. The findings are validated by an implementation of the proposed algorithm(s) and comparison agains reasonable baselines.
Students wanting to work on this topic are expected to have solid mathematical skills, a basic understanding of machine learning techniques, solid knowledge of Python and deep learning libraries (PyTorch or TensorFlow).
Contact: Sebastian Tschiatschek
Posterior Consistency in Partial Variational Autoencoders
Variational Autoencoders (VAEs) are powerful deep generative models that have been successfully applied in a wide range of machine learning applications. Recently, the Partial VAE (PVAE), a variant of VAEs that can process partially observed inputs has been proposed and its effectiveness for data imputation has been demonstrated. Key to the fast training of VAEs and PVAEs is the amortized prediction of posterior distributions from observations. In PVAEs, these posterior distributions are predicted from partial observations.
This project aims at studying the consistency of these posterior distributions for different patterns of missing data. The insights are used to create/train better inference models and thereby improve the quality of PVAEs.
Students wanting to work on this topic are expected to have solid mathematical skills, a basic understanding of machine learning techniques and good programming skills in Python.
Contact: Sebastian Tschiatschek
Evaluation of different nowcasting techniques and data bases for temperature nowcasting
Nowcasting in meteorology refers to short-time prediction of timeseries of e.g. a few minutes to hours into the future. For this short-time predictions the most recent observations are of importance as well as spatial relationships of upstream observations.
Different methods for nowcasting exist in the meteorological world, often physics of physics-statistical driven. Only a few studies focused on applying machine learning and data mining techniques. For the latter nowadays also crowd-sourced information can contain important information. Here, data of NetAtmo stations could provide valuable information. For the first part a study investigated already different machine learning algorithms for nowcasting of temperature using the Austrian Met-Service (ZAMG) data. The idea of the proposed thesis is to combine the already existing algorithms with the NetAtmo data by first clustering the ZAMG data and NetAtmo data and then develop an algorithm which incorporates the NetAtmo data. Important to note is that NetAtmo sites are prone to errors thus need to be cleaned beforehand.
The developed algorithm will be evaluated against two statistical forecasting methods, an analogue search based method and a model output statistics method.
Contact: Irene Schicker, Claudia Plant
Causal inference among climatological time series with extreme events [Master thesis]
This work will focus mainly on implementation of an algorithm for causal detection among processes as well as its testing on probability distributions with heavy tailed distributions and on climatological data provided by ZAMG (Die Zentralanstalt für Meteorologie und Geodynamik) following these distributions. More details to the interested student will be provided on request.
Contact: Katerina Schindlerova
Implementation of Data Mining Approach for Short-range Temperature Forecasts
Short-range forecasts of wind speeds (i.e., 1 - 72 hours into the future) and in particular nowcasting (i.e., very short-range forecasts with a time horizon of up to 6 hours) are vital for a wide range of applications. In contrast to wind, temperature typically changes gradually and may thus need a different setup than wind speed. Temperature involves daily fluctuations, which are well predictable but dependable on the daytime, which must be considered in the training dataset. Depending on the location, also rapid temperature changes may occur which are related, for example, to cold air pools or Föhn. Thus, also temperature highly depends on location (specific topography, prevailing weather conditions, and atmospheric dynamics). Depending on the application, we can either give point or area based predictions. Points refer to a particular location (e.g., a weather station) whereas spatial forecasts typically give a forecast for each grid cell over a region (i.e., each forecast is valid for the whole cell).
ZAMG (Zentralanstalt für Meteorologie und Geodynamik) employs (gridded) numerical weather prediction models in conjunction with observation data for short-range forecasting and a nowcasting system, INCA, for the prediction of meteorological parameters. Alternatively, machine learning methods are now being implemented. In particular, an artificial neural network (ANN) and a random forests (RF) are used in an experimental setup to show the skills of these methods and, possibly, be used as additional wind speed point forecasting method for the 10-meter wind. The existing methods can be used for temperature as well with the same training set, but the setup needs to be adapted.
The proposed student project shall address temperature forecasting (in two meters height) at meteorological observation sites by machine learning and data mining methods (e.g.: random forest, feed forward/BP artificial neural networks, kernel methods, etc.) and input feature selection of the training data set. It is possible to experiment with related meteorological parameters as well (e.g.: drew point temperature and relative humidity). Related work suggests to use back propagation neural network for predicting the 2-m drew point temperature and 2-m temperature. This can be used as a starting point to find a suitable selection of the training data for the current model and then extend the our current approaches by a back propagation neural network in order to set up a first prototype for temperature prediction by machine learning methods. The new model shall be tested on various scenarios (e.g.: different prevailing weather condition, locations, seasons) in order to compare the new data mining based model with the currently employed nowcasting system INCA.
The work is co-supervised by ZAMG's Section for numerical weather predictions (NWP) Applications. The developed method shall have a Python based frontend, C/C++ backend, and use csv or sqlite based meteorological data (provided by ZAMG) in order to align with other machine learning implementations running in our IT environment. Finally, the developed method will be set up in our development environment (Python 2.7/Linux 64-bit, multi-cored shared memory machine) to provide forecasts and validation of the method for selected test scenarios.
Contact: Petrina Papazek, Claudia Plant
Completed
- Wei Chen, Bachelor Thesis: "Mining Brain Networks", winter term 2022/2023
- Kejsi Hoxhallari, Bachelor Thesis: "Statistical validation and visualization of causal inference with extremes in wind-turbine data set", winter term 2022/2023
- Daan Scheepens, Master Thesis: "A deep convolutional RNN model for spatio-temporal prediction of wind speed extremes in the short-to-medium range for wind energy applications", winter term 2021/2022
- Yigit Berkay Bozkurt, Bachelor Thesis: "Anomaly Detection by Heterogenous Graphical Granger Causality and its Application to Climate Data", 2019
- Christina Pacher, Bachelor Thesis: "Clustering Weather Stations: A Clustering Application for Meteorological Data", summer term 2019
- Thomas Spendlhofer, Bachelor Thesis: "Evaluating the usage of Tensor Processing Units (TPUs) for unsupervised learning on the example of the k-means algorithm", summer term 2019
- Ernst Naschenweng, Bachelor Thesis: "A cache optimized implementation of the Floyd-Warshall Algorithm", summer term 2018
- Hermann Hinterhauser, Bachelor Thesis: "ITGC: Information-theoretic grid-based clustering", summer term 2018, accepted paper in EDBT 2019 (download available here)
- Mahmoud A. Ibrahim, Bachelor Thesis: "Parameter Free Mixed-Type Density-Based Clustering", winter term 2017/2018, accepted paper in DEXA 2018 (download available here)
- Markus Tschlatscher: "Space-Filling Curves for Cache Efficient LU Decomposition", winter term 2017/2018
- Theresa Fruhwuerth, Master Thesis: "Uncovering High Resolution Mass Spectrometry Patterns through Audio Fingerprinting and Periodicity Mining Algorithms: An Exploratory Analysis", summer term 2017
- Robert Fritze, PR1 "Combining spatial information and optimization for locating emergency medical service stations: A case study for Lower Austria", summer term 2017
- Alexander Pfundner, PR2 "Integration of Density-based and Partitioning-based Clustering Methods", summer term 2017
- Anton Kovác, Katerina Hlavácková-Schindler, Erasmus project, "Graphical Granger Causality for Detection Temporal Anomalies in EEG Data", winter term 2016/2017 (download available here)