DISA

Centre for Data Intensive Sciences and Applications

Welcome to our PhD-seminar in February

Postat den 2nd February, 2026, 13:03 av Elin Gunnarsson

Where? Onsite: D2272 and via zoom
Registration: Please sign up for the seminar via this link https://forms.gle/QB5oiSWMpVofjBHY6 by February 11. This is especially important if you plan to attend onsite so we can make sure there is fika for everyone.

Abstracts
Novelty Detection Using Time-Series Transient Data from the Lead-Acid Forklift Batteries – Zijie Feng  
In industrial applications, monitoring the battery health of electrically powered machinery and detecting abnormal operating conditions is a persistent yet critical challenge. Traditionally, anomaly detection systems are developed reactively: abnormalities are identified only after they occur, often leading to operational disruptions and economic losses. Novelty detection offers an alternative by learning normal behavior and detecting previously unseen abnormalities.  

In this work, we compare a diverse set of data-driven novelty detection methods using simulated time-series transient data derived from real lead-acid forklift battery measurements, aiming to identify suitable solutions for different types of anomalies. 

Potential of Graph Neural Networks for Software Architecture Recovery– Rakshanda Jabeen  
Software architecture recovery (SAR) aims to uncover a system’s modular structure directly from source code, supporting comprehension and maintenance when documentation is missing or outdated. In this work, we investigate the potential of graph neural networks (GNNs) and unsupervised learning for SAR by modeling software systems as heterogeneous, multi-relational graphs. Nodes represent software entities (e.g., files or classes) and typed edges capture structural and functional dependencies such as calls, imports, inheritance, and other code-level relations. To complement dependency structures with meaning, we integrate semantic signals from source code identifiers and related textual artifacts via contextual code embeddings (e.g., Word2Vec), yielding representations that capture both what entities do and how they interact. 

We study heterogeneous GNN encoders that aggregate information across relation types, including heterogeneous graph convolution and heterogeneous attention mechanisms, to analyze the trade-off between fixed normalization and adaptive neighbor weighting in software graphs. On top of these encoders, we explore two unsupervised training paradigms: (i) graph autoencoding, where embeddings are learned by reconstructing observed dependency relations, and (ii) contrastive representation learning inspired by Deep Graph Infomax, which maximizes agreement between embeddings from the original graph and perturbed views. The resulting entity embeddings are clustered to recover candidate architectural modules. Preliminary results across multiple open-source systems indicate that combining semantic cues with structural and functional dependencies produces more meaningful module separation than using structure alone, demonstrating that modern graph representation learning is a promising direction for robust, automated SAR beyond heuristic baselines. 

Det här inlägget postades den February 2nd, 2026, 13:03 och fylls under General

Comments are closed.