Hide menu

Examensarbeten och uppsatser / Final Theses

Framläggningar på IDA / Presentations at IDA


Se även framläggningar annonserade hos ISY och ITN i Norrköping / See also presentations announced at ISY and ITN in Norrköping (in Swedish)

If nothing is stated about the presentation language then the presentation is in Swedish.


WExUpp - kommande framläggningar
2024-05-20 - HCS
Damage Assessment on Remote Sensing Imagery with Foundation Models
Gustaf Lindgren
Avancerad (30hp)
kl 08:15, Alan Turing (In English)
[Abstract]
There is currently an ongoing paradigm shift in machine learning; instead of training task-specific models from scratch, foundation models i.e., large pre-trained models are adapted for various downstream tasks. Foundation models excel in zero- and few-shot learning, ideal for domains with limited labeled data, such as disaster assessment on remote sensing imagery (RSI).

This thesis explores how the foundation models CLIP and SAM can be utilized to classify RSI affected by natural disasters and segment intact and damaged infrastructure without extensive retraining. For the scene classifications, various text prompt techniques are tested as well as zero-shot prompting with images. Moreover, few-shot learning methods such as linear probing and prompt learning are explored. For the open vocabulary semantic segmentation task, "pipelines" are implemented that leverage the open vocabulary classification abilities of CLIP and zero-shot image segmentation capabilities of SAM.

This work demonstrates that foundation models can be used effectively for detecting flooding on RSI and there were promising results on other disaster types as well. While handcrafted text prompts yielded the best accuracy, the zero- and few-shot learning methods with images offered a better trade-off between accuracy and consistency. Although the performance of the zero-shot segmentation pipelines was generally poor, they showcased the potential of SAM for accurate segmentations on disaster imagery when being provided with prompts of sufficient quality.
2024-05-20 - AIICS
Forecasting Patient Occupancy in Hospital Wards Using a Supervised Machine Learning Approach
Axel Falk, Philip Folkunger
Avancerad (30hp)
kl 09:15, John von Neumann (In English)
[Abstract]
o The healthcare sector faces challenges in balancing resource allocation and meeting patient demand, especially in the Emergency Department (ED) and other wards. This study explores the potential of supervised machine learning models to predict occupancy rates across different hospital departments using data from a hospital in western Norway from 2020 to 2023. The research combines Fourier analysis, seasonal decomposition (STL), and crosscorrelation techniques to identify cyclical patterns and dependencies within the data. Various supervised machine learning models, including Linear Regression, Random Forests, XGBoost, and neural networks, are evaluated using k-fold cross-validation and performance metrics such as MAPE and MAE. The results reveal distinct daily and weekly patterns in hospital occupancy rates, with notable anomalies during holidays and weekends. The study finds that occupancy rates are consistent over time, as ED, Cardiology Ward (CW), and Total Patients (TP) series are stationary, with stable mean values and variances. Both TP and ED exhibit daily seasonality, while all three series display weekly seasonality. Machine learning models perform differently across wards. The smallest prediction errors using only time features were 5.595 MAE for ED, 1.794 MAE for cardio, and 0.096 MAPE for TP. Cross-correlation analysis revealed strong correlations in daily cycles between ED and TP when lagged in time, suggesting that ED and TP occupancy rates are closely linked, while cardio shows slightly different patterns. The study concludes that simpler models, like Linear Regression, may offer a more efficient and effective approach for hospital occupancy forecasting.
2024-05-20 - AIICS
Large Language Models as Outlier Detectors - How Large Language Models can be used to detect outlying datapoints in business data
Oskar Gunnarsson, Joel Melkersson Dalén
Avancerad (30hp)
kl 10:15, Alan Turing (In English)
[Abstract]
The data available to companies is ever-growing, making value-extraction from this data a differentiating factor. This thesis studies how Large Language Models can be used to extract outlying datapoints from business data, with three different prompts as input. The performance of the Large Language Model was compared to that of traditional Machine Learning models. ChatGPT was used for the Large Language Model and KNN, ECOD, PCA, and iForest were the traditional models used. The Large Language Model successfully found some outlying datapoints and did so with similar proficiency to the traditional models. However, none of the models performed great. Although the Large Language Model's results were not exceptional, they showed potential for improvement with use of task-specific alterations. An area where these improvements could make Large Language Models outperform the traditional models is in the historically difficult task of detecting contextual anomalies. Further research using task-adapted prompts, a domain-specific training corpus, and evaluation with a variety of datasets is needed to fully gauge the possibility of using Large Language Models as outlier detectors in a business context.
2024-05-20 - AIICS
LLM for Natural Language Domain Description
Elliot Gestrin
Avancerad (30hp)
kl 14:15, Alan Turing (In English)
[Abstract]
Today's classical planners are powerful, but modeling input tasks in formats such as PDDL is tedious and error-prone. In contrast, planning with Large Language Models (LLMs) allows for almost any input text, but offers no guarantees on plan quality or even soundness. In an attempt to merge the best of these two approaches, some work has begun to use LLMs to automate parts of the PDDL creation process. However, these methods still require various degrees of expert input. We present NL2Plan, the first domain-agnostic offline LLM-driven planning system. NL2Plan uses an LLM to incrementally extract the necessary information from a short text prompt before creating a complete PDDL description of both the domain and the problem, which is finally solved by a classical planner. We evaluate NL2Plan on four planning domains and find that it solves 10 out of 15 tasks - a clear improvement over a plain chain-of-thought reasoning LLM approach, which only solves 2 tasks. Moreover, in two out of the five failure cases, instead of returning an invalid plan, NL2Plan reports that it failed to solve the task. In addition to using NL2Plan in end-to-end mode, users can inspect and correct all of its intermediate results, such as the PDDL representation, increasing explainability and making it an assistive tool for PDDL creation.
2024-05-22 - HCS
Säkerställande av förarsäkerhet vid interaktion med touchskärmar för arbetsverktyg i bilar
Johanna Lundin
Avancerad (30hp)
kl 10:15, IDA Alan Turing (På svenska)
[Abstract]
The integration of devices within cars is continuously evolving, enabling us to interact with them to an increasingly greater extent. This has transformed the way we drive, communicate and access information on the go. Despite this, there is a lack of research on how to guarantee driver safety while interacting with these systems, especially when looking at in-car systems used in professional settings as a work tool. This master's thesis was conducted in collaboration with NIRA Dynamics and aimed to investigate how the interface of in-car touch screen work tools can be designed to ensure usability and safety for the driver. The study included development of a prototype in the form of a new touch screen interface for a data acquisition system which was used by NIRAs test drivers for the purpose of testing their products. The prototype design was developed iteratively based on the test drivers' opinions as well as theory about important design aspects related to designing in-vehicle systems for high safety and usability. The resulting prototype was evaluated using the System Usability Scale in order to compare the prototype to the original system design and asses to what extent the new interface contributed to increased safety for the driver. The study revealed that some of the main issues that needed to be taken into account in the prototype design was prioritization of information, placement and gathering of elements and reduced amount clicks and scrolling. The final usability evaluation was conducted through user tests and the results indicated that the usability of the prototype was higher than for the original system design, thereby indicating an increased safety for the driver. Overall, this thesis contributes to the research of mitigating the risks of drivers related to interaction with in-car software systems.
2024-05-23 - HCS
Bemästring och rörelse över tid i fysikbaserade 2D träningsspel
Filip Josefsson, Jim Magnusson
Avancerad (30hp)
kl 13:00, Donald Knuth (På svenska)
[Abstract]
In the modern world, people are not moving enough. People can sit down for multiple hours in a row at an office job, which can lead to multiple health risks like obesity, depression and cancer. One way of counteracting this sedentary behavior is to encourage people to move more by playing exergames, video games that require the player to move and exert themselves to play. The benefit of playing exergames does, however, only apply when they are continually played over a longer period of time. To see how well exergames work over a longer period of time, two casual physics based exergames were developed. Each exergame was then played 100 times each by both authors to see how score and movement changed over time. It was found that both games required greater movement to get higher scores. The movement amount of the exergames slightly decreased over time in some cases but, most often, stayed at the same level or increased over time.
2024-05-23 - SaS-UND
Energy consumption of video streaming – A literature review and a model
John Lindström
Avancerad (30hp)
kl 13:00, Alan Turing (In English)
[Abstract]
Energy consumption and correlated greenhouse gas emissions are a big global problem.
It affects all parts of society, and each industrial sector must work toward reducing its
carbon footprint. This thesis details the research of different methods to model the energy
consumption of video streaming, and works towards creating a final model. The video
streaming process is broken down into a core process consisting of head-end, distribution
and transmission, and terminals. The process that contributes the most to energy consumption
at the head-end is found to be video encoding. This thesis explores video encoding in
depth and how it is affected by parameters such as hardware, codec choice, codec preset
selection, and video details such as resolution, framerate, and duration, but these parameters
are found to be insufficient to model the energy consumption of video encoding. In
distribution and transmission, the highest contributor is found to be content delivery networks.
The energy consumption of content delivery networks is investigated however no
appropriate model is found. For terminals, the most important factor is the kind of terminal
used. The energy consumption of televisions, desktop computers, laptops, and mobile
terminals is investigated, and models are presented for each. The thesis also discusses the
different models, their advantages, and their shortcomings. Additionally, an application to
visualize features of the model is created and presented.
2024-05-24 - HCS
Utforskande av exergaming: En studie om spelarbeteenden och speldesign
Matthias Gerdin, Lisa Green
Avancerad (30hp)
kl 09:00, IDA Alan Turing (På svenska)
[Abstract]
This thesis investigates the potential use of exergames, games controlled by the player’s body movements, to promote large and powerful arm movements as a means to combat sedentary behavior. Two score-driven games, Smash It and Pop It, are developed to explore whether a scoring system could incentivise players to maintain or increase their arm movements as their skill levels increases. The study employ a methodology involving data collection from participants engaging with the games, with a focus on analyzing movement and gameplay performance. Results indicate promising outcomes, with both games demonstrating effectiveness in motivating desired movements. While Smash It offers players freedom in gameplay, Pop It provides a more linear experience, allowing developers greater control over player actions.
2024-05-24 - HCS
The Impact of Level Design on Movement in a 2D Platformer Exergame
Josef Karlsson
Avancerad (30hp)
kl 13:15, Kurt Gödel (På svenska)
[Abstract]
This thesis investigates how level design impacts user movement in a 2D platformer exergame and how this movement evolves as players gain experience. A single participant played the exergame 100 times, with 30 minutes of rest separating each session. Findings indicate that level designs that require or encourage waiting behavior reduce user movement and limit the potential for increased movement as the user gains experience.
2024-05-31 - AIICS
Strategies for Accurate Context Retrieval in Retrieval-Augmented Generation Systems Across Diverse Datasets
Axel Andersson, Hugo Björk
Avancerad (30hp)
kl 13:15, John von Neumann (In English)
2024-06-04 - ADIT
Evaluation of Unsupervised Anomaly Detection in Structured API Logs
Gabriel Hult
Avancerad (30hp)
kl 13:15, Charles Babbage (In English)
[Abstract]
With large quantities of API logs being stored, it becomes difficult to manually inspect them and determine whether the requests are benign or anomalies, indicating incorrect access to an application or perhaps actions with malicious intent. Today, companies can rely on third-party penetration testers who occasionally attempt various techniques to find vulnerabilities in software applications. However, to be a self-sustainable company, implementing a system capable of detecting abnormal traffic, potentially maliciously, would be beneficial. By doing so, attacks can be proactively prevented, mitigating risks faster than waiting for third parties to detect these issues. A potential solution to this is applying machine learning, specifically anomaly detection, which is detecting patterns not behaving to the normal standard. This thesis covers the process of having structured log data to find anomalies in the log data. Various unsupervised anomaly detection models were evaluated on their capabilities of detecting anomalies in API logs. These models were K-means, Gaussian Mixture Model, Isolation Forest and One-Class Support Vector Machine. The findings from the evaluation show that the best baseline model without tuning can reach a precision of 63%, a recall of 72%, resulting in an F1-score of 0.67, an AUC score of 0.76 and an accuracy of 0.71. By tuning the models, the best model could reach a precision of 67% and a recall of 80%, resulting in an F1-score of 0.73, an AUC score of 0.83 and an accuracy of 0.75. The pros and cons of each model are presented and discussed along with insights related to anomaly detection and its applicability in API log analysis and API security.
2024-06-05 - HCS
Automating Figma Design to Web Component Conversion: Enhancing Workflow Efficiency in Web Development
Oliver Börjesson
Avancerad (30hp)
kl 10:00, Kurt Gödel (In English)
[Abstract]
This thesis aims to address the significant challenges in the collaboration between designers and developers in web and application development. The focus is on developing a tool or methodology that bridges the gap between design and implementation, thereby creating a more seamless, efficient, and error-free workflow. Existing methodologies and tools often favor either designers or developers and are time-consuming, leading to repeated revisions and misunderstandings. This research proposes a design-to-code conversion tool that streamlines the process, reduces errors, and enhances productivity. Through a comprehensive literature review, prototype development, and empirical evaluation, this study explores the technical requirements, best practices, and user feedback mechanisms necessary for creating an effective tool. The results demonstrate that while the prototype offers promising improvements in design-to-code translation, further refinements are necessary to address dynamic positioning and broader design tool compatibility. The study concludes with recommendations for future work, including support for more complex design elements and the integration of machine learning techniques to enhance the tool's capabilities.
2024-06-05 - ADIT
How does the use of Autonomous Penetration Testing Strengthen The Continuous Integration Flow?
Jonatan Eshak
Avancerad (30hp)
kl 13:15, Herbert Simon (In English)
[Abstract]
The thesis introduces the problems developers face when creating, optimizing, and testing their systems. The focus of this thesis is the testing of a web application with the use of autonomous penetration testing integrated into a GitLab CI/CD pipeline. The thesis wants to answer whether the use of open API, a specification made to ease the documentation of API in a system, creates an environment where one can save time on integration and increase efficiency through knowledge of performance from specific endpoints. The thesis wants to measure further the applications where autonomous penetration testing with open API could be preferred. The thesis also measures the different use cases of using autonomous black box testing against a white box to answer when one can be preferred over the other and when it is helpful to have both. The thesis goes through the theory of penetration testing and how it is conducted, what common strategies, and what attacking methods are standard. The thesis also goes into the theory of autonomous PT against manual PT, a theory of web application, and API to describe open API and what Swagger is as a tool. The thesis also goes into the theory of continuous integration flow, its design, and how a developer builds one from scratch. The thesis also brings up five significant articles related to the work, such as an article that discusses the problems faced when designing a black box vulnerability scanner on web servers. This article discusses the implementations of continuous integration on automatic performance testing and one for automating security scanning. The thesis also brings up an article on introducing continuous fuzzing to study the scalability of fuzzing in CI/CD pipelines. These related works enhance the purpose and method of this thesis and its goal to measure autonomous penetration testing in a CI pipeline. The method to answer the research question is to build a website to serve as a target with open API and integrate the website into the GitLabs CI pipeline along with vulnerability scanning tools configured to perform black box, grey box, and white box testing. The results show that while the black box is a lot more thorough through having to search and test every different point of discovery, it does so at the cost of time. The grey box shows similar results as the black box, although it only focuses on finding vulnerabilities from API endpoints. Using the white box, the results showed more critical vulnerabilities mainly focused on packages installed and stored in an environment directory. The vulnerabilities also differentiate from the black box and grey box, showing a need to use both scans to discover as many unique vulnerabilities as possible.
2024-06-17 - ADIT
Characterizing Player Types in Ice Hockey: A Data-Driven Approach
Anton Olivestam, Axel Rosendahl
Avancerad (30hp)
kl 13:15, Charles Babbage (In English)
[Abstract]
Ice hockey is a team sport, and thus it is of the utmost importance to assemble a team of players who are capable of performing at their optimal level when playing together. For a team to perform at the highest level, it is essential that the team consists of players with different player types who excel in different situations on the ice. Despite the importance of team composition, there is a shortage of previous research in this area. Previous studies have been limited to a single league or the quantity of available data. This thesis investigates the classification of player types in ice hockey using soft clustering techniques. The aim is to determine if fuzzy clustering and Gaussian mixture models (GMM) can effectively categorize player styles, and to evaluate the suitability of each method. The player types were derived from play-by-play data from three different leagues over three seasons. We represent teach player's individual playing style in two different player vectors. One vector employs frequent sequences combined with event frequency, while the other vector features distinct skills characterizing a playing style. After constructing the vectors, we applied both of the clustering algorithms. Our findings show that both FCM and GMM successfully derived playing styles. FCM provided fuzzier clusters, meaning that it performed better at handling the fact that players may assume multiple playing styles. Additionally, the vectors based on the distinct skills of a hockey player led to highly interpretable clusters compared to the vectors using frequent sequences and event frequency.
2024-06-18 - ADIT
Run-Time Optimization of ElasticSearch for Medical Documents
Ludvig Bolin, Emil Carlsson
Avancerad (30hp)
kl 13:15, Babbage (In English)
[Abstract]
ElasticSearch is a database management system used to index and search documents, and as with all database management systems, performance is important. The aim of this thesis is to investigate whether the configuration of an ElasticSearch system can be tuned to improve either index- or search performance using different optimization algorithms. With that goal in mind, this thesis has evaluated three different optimization algorithms as a means to generate performance-improving ElasticSearch configurations. Two local algorithms, Simulated Annealing and Simultaneous Perturbation Stochastic Approximation, and one global algorithm, Genetic Algorithm.
The benchmarking tool ESRally is used as an objective function for the local algorithms. Since the global algorithm requires near-instant evaluation, two machine-learning models are trained to predict configuration performance in said benchmarks instead. The machine learning models, Random Forest, and Regression-Enhanced Random Forest, performed with similar accuracy. Both models could predict the performance of a configuration measuring index performance well but could not predict the search performance to the same extent.
The configuration generated by the various optimization algorithms is then evaluated in a simulation replaying four hours of real traffic from an ElasticSearch instance used in a hospital for medical data indexing and searching. Unfortunately most configurations generated by the various algorithms failed to improve search performance. On the other hand all the algorithms succeeded in generating configurations that outperform the default configuration in the simulation regarding indexing performance, with Simultaneous Perturbation Stochastic Approximation producing the best performance configuration.


Page responsible: Final Thesis Coordinator
Last updated: 2022-06-03