Master Thesis
Evaluation Guidelines
The master thesis project is evaluated by a committee constituted by three members: the thesis supervisor and two other members of the department. This evaluation considers the tutoring sessions, the oral presentation and the written report.
Important dates
Early Jan | Agree with your supervisor on a topic |
31 March | First draft containing motivation, context, state of the art and scientific questions. |
Late June | Submission of the draft report to the panel committee. |
July | Oral presentations and submission of the final report to the panel committee. |
Oral presentation
Oral presentations of the thesis will be scheduled in mid July. They will take 20 minutes plus 10 minutes of questions from the evaluation committee.
Thesis report
The student has to submit a draft version of the thesis report to the evaluation committee before June 30th. Accordingly, the student and the supervisor have to arrange that the supervisor has enough time to read and assess the final thesis text before the presentation. The final version of the report must be submitted to the panel committee before the oral presentation.
Formatting of the report
As for the format of the written thesis (Font size, line spacing, margins, Section numbering etc) we propose that the students follow the template provided by the Universitat Pompeu Fabra for master theses. For Latex you can use this one. The MIIS board does not have any preference on the word processor and will accept documents written with any word processor.
The actual report can be structured according to Introduction, State of the Art, Methods, Results, Discussion and Conclusion. However, this is not mandatory. Different overall organizations of the report can be used, if necessary. As a general guideline we would like to indicate that a Master report is not supposed to reach the comprehensiveness of a PhD thesis. An adequate length of the thesis is between 30-50 pages (this number is regarded as a guideline).
More Information
Proposed thesis topics
Title & Description |
Supervisor(s) |
Self-supevised multi-modal 2D and 3D image registration Image registration refers to the problem of aligning two or more images having similar content, transformed by an unknown geometric transformation (a rotation, an homography, or a more complex displacement field). In multi-modal image registration the images to be registered correspond to different modalities. Examples are aligning an RGB image with a thermal image, or a computer tomography scan with a magnetic resonance image. Multi-modal image registration is a challenging long standing problem with several applications in computer vision and medical image analysis. Neural networks have demonstrated great potential for image registration, yet they require large training datasets where the ground truth transformation between the images has to be known. Unsupervised training methods are appealing as they enable training when the ground truth transformations are unknown. Several unsupervised methods have been proposed for registering images of the same modality, but unsupervised multi-modal registration has remained largely unexplored. Recently, a fully unsupervised registration method has been proposed in the context of multi-spectral satellite images (see link below). The goal in this thesis is to extend this method for its application in 2D and 3D medical imaging. |
|
Efficient policies for constrained reinforcement learning In this project, we will study decision-making algorithms under requirements. Many practical applications make use of learning algorithms which are required to satisfy constraints such as safety, energy use or robustness. Under the reinforcement learning framework, it is the agent who learns to satisfy these by interacting with the environment, in what is known as constrained reinforcement learning. We will study constrained reinforcement learning formulations and exploit recent developments in duality theory to improve on current solutions, designing efficient and versatile methods capable of guaranteeing near-optimality and minimal constraint violation. |
|
Task-uncertain reinforcement learning In this project we will study the problem of designing reinforcement learning policies in settings where we do not know beforehand which task must be performed. We will consider reinforcement learning problems with different objectives (rewards), each describing a task that may need to be performed during deployment (e.g., deviating from a large or small obstacle). At execution time, the agent must decide which task to perform by interacting with the environment and collecting rewards. In other words, in this project, we will design and test algorithms exploiting rewards not only to learn how to perform tasks, but also to identify which task to perform. |
|
Master Thesis Proposals at Eurecat technology centre This year, the Robotics and Automation Unit at Eurecat is offering different proposals to be developed as a Master Thesis, which are listed below. If the student has its own proposal, it also could be taken it in consideration. |
|
Heuristic design for dynamic feature selection In some situations, information usage must be minimized, for instance in a medical diagnostic setting, where acquisitions can be invasive or dangerous for the patient. Reinforcement learning allows to optimally solve this problem, proposing at each step the modalities that will minimize the information/cost balance based on the currently acquired data not only greedily, but also accounting for future decisions. However, exploring the search space of all possible features combinations is computationally unfeasible. In this thesis, you will explore heuristics to quickly discard regions of the space that are not relevant due to their low value. In particular, you will derive confidence bounds for sampling-based heuristics that will allow to balance the exploration-exploitation trade-off. |
|
Algorithmic hiring This is a project in the algorithmic fairness space, and is part of Horizon Europe project FINDHR. Algorithmic hiring is the usage of tools based on Artificial intelligence (AI) for finding and selecting job candidates. As other applications of AI, it is vulnerable to perpetuate discrimination. There are various subprojects that can be done here, including generation of synthetic data, study of ranking and recommendation algorithms, and external audits of actual hiring platforms. |
|
Gender bias in online platforms This project will analyze and study potential gender biases of online platforms. These may include job search engines, e-commerce websites, search engines, and platform economy sites for which some data could be scrapped/obtained. The specific platform is to be decided with the master student. The work will involve data scrapping/collection, measurement through fair ranking/recommendation algorithms, and analysis of potential biases found. It requires solid knowledge of Python and desire to work within a framework of data feminism, which seeks to eliminate gender biases in data collection and data processing pipelines. |
|
Risk-Prediction Instruments in Criminal Justice The Justice Department of Catalonia has been using for several years structured risk prediction instruments, similar to COMPAS in the US or OASIS in the UK, to predict the risk of recidivism and violent recidivism. Correctional officers, psychologists, and social workers apply a questionnaire to juvenile offenders and to people who have been convicted for a crime (these are different questionnaires); the questionnaire is then processed using a model trained on thousands of previous cases, and an output is produced, which is then interpreted by the person applying the instrument. Detailed anonymized data is available for studying to what extent different instruments can be biased against different sub-populations (e.g., immigrants or children of immigrant). Additionally, we would like to explore the potential for new risk prediction instruments. The work would be done in collaboration with researchers on criminology at the University of Barcelona and at the Justice Department of Catalonia. |
|
Chat with your documents, table and images with Retrieval Augmented LLMs Large language models (LLMs) and generative AI have taken the world by storm thanks to their incredible capabilities of fluency and knowledge extraction. However, often information stored in the model during training is obsolete or not relevant to the current task. One exciting possibility is to use a vector database as a memory and to let the language model access this information at inference time. This approach results in answers that are more factual answers and relevant. In this project, we will jointly investigate how to develop and validate a Retrieval Augmented system that uses LLMs to create embedding’s for multimodal documents, including text, table and images from public literature and technical reports from the European Commission. |
|
Multicentric clinical study on lung cancer Lung cancer is one of the most deadly diseases in the world and recent results showed that early detection by means of an yearly Computer Tomography (CT) scan of the chest reduced mortality by 20% in a population of ex-smokers. However, once a lesion is found in the CT scan, then a biopsy is needed to confirm whether the nodule is malignant, which is a very invasive procedure. The JRC participates in a clinical study with 4 hospitals to enable and advance the use of AI in the detection and follow up of lung lesions. In the scope of this project, we will jointly work on the cloud infrastructure that processes the clinical data and on the deep learning algorithms for the analysis of the CT scans. |
|
Reasoning with language and knowledge graphs Large language models (LLMs), are making new waves in the field of natural language processing and artificial intelligence, due to their emergent ability and generalizability. However, LLMs are black-box models, which often fall short of capturing and accessing factual knowledge. In contrast, Knowledge Graphs (KGs), Wikipedia and Firebase for example, are structured knowledge models that explicitly store rich factual knowledge. KGs can enhance LLMs by providing external knowledge for inference and interpretability. Meanwhile, KGs are difficult to construct and evolving by nature, which challenges the existing methods in KGs to generate new facts and represent unseen knowledge. Therefore, it is complementary to unify LLMs and KGs together and simultaneously leverage their advantages. In this project, we will work jointly towards a system that unifies LLM and KGs for the processing of scientific literature and technical reports from the European Commission. |
|
Genomic similarities via vector embeddings Recent advances in sequencing techniques created a huge amount of public genomic data available for viruses, bacteria and mammals. That wealth of data has the potential to unlock important advances in all the fields of medicine such as communicable and non-communicable disease, pharmaceutical and antimicrobial resistance. However, processing those data in an efficient way and creating models that can generate recommendations based on them is still a challenge. In this project, we will work on deep learning methods to process genomic and molecular data to create vector embeddings that can then be used for fast similarity search. |
|
Transferable Adversarial Attacks on Aligned Language Models for Health Despite the success of deep neural networks, the existence of adversarial attacks has revealed the vulnerability of neural networks in terms of security. Adversarial attacks add subtle noise to the original example, resulting in a false prediction. Although adversarial attacks have been mainly studied in the image and speech domain, a recent line of research has discovered that LLMs are also exposed to adversarial attacks. This is now starting to appear also in the multimodality of Large Language Models. In this project we aimed, working with LLMs and making use of the research literature available, to propose multimodal adversarial attacks that can induce aligned language models to produce questionable content which does not follow the HHH principles (helpful, honest, and harmless) in the models. |
|
Data Analysis in Education Technologies Several topics related to the application of AI and data analytics techniques to the design of systems that support teaching and learning (includes community, teaching and learning analytics, interactive dashboards, etc.) |
|
Urban data governance. Beyond open data How data from public sector, private companies and citizens can be managed and governed to extract more richness while keeping privacy and securing data explotation models. The EU’s Data Governance Act seeks for promoting data reuse and the GAIA-X project is meant to provide an architecture for such ideal. Current debates around Data Spaces and Data Trusts, some include a Federated infraestructure and some others trust in DAOs (blockchain) to provide a solution that includes democratic approach to data sharing. This TFM will explore different opportunities and question what are the assumptions behind each model and understand how different infraestructure is needed to accomplish the needs of different uses of data in the urban context. It will be focused on theoretical analysis of different models and technical assessments of current data technologies and architectures. In addition, the analysis of different spatial-temporal datasets will be analysed to study how cities can integrate and benefit from this approach. |
|
Intersectional Fairness During the last years, Fairness Accountability and Transparency (FaccT) has been discussed to reduce social harm and discrimination in the design and development of algorithms. Although, most of the works on FaccT approaches only address discrimination towards limited categories in just one dimension (e.g., gender, race). Recently, voices from black feminism and social studies have called to a more depth approach to a fair and just development. Intersectionality is an analytical framework for understanding how aspects of a person's social and political identities combine to create different modes of discrimination and privilege. It identifies multiple factors of advantage and disadvantage. Examples of these factors include gender, caste, sex, race, class, sexuality, religion, disability, physical appearance, and height. That is, it goes behind the classical gender/race binaries. Thus, looking at one dimension at a time doesn’t always tell us the whole story. Despite that the call for an intersectional approach is growing among the Computer Science and Data Science communities, few works explore the real possibility of working the concept of FAccT over algorithms. The goal of this thesis is to develop an exploration of diverse datasets and algorithms used in the context of social interest (e.g., social welfare, recidivism, education) and to develop technical solutions taking into account an intersectional perspective. This thesis is approached with a highly theoretical conceptualization as well as a technical development. |
|
Fair housing algorithm Housing market in Barcelona is under constant scrutiny due to high demand and current economic-crisis. Moreover, the housing market prices are determined by the same longstanding valuations without reflecting current societal and environmental needs (e.g. green areas, low pollution, and diversity). The goal of this thesis is to understand which factors related to climate change and pollution could affect the real state market and housing policies in the near future. The project is meant to analyze different sources of data and to analyze the current status of Barcelona’s housing market from a data-driven approach. As a result, it is expected to develop an algorithm to re-calculate the cost of housing, not under market demand-offer dynamics, but under Sustainable Development Goals (SGDs) indicators, the impact of climate change and social justice principles. |
|
Ranking Attributes for Information Gain in Forests of Decision Trees (data science/machine learning related project) Decision trees are one of the most well-known techniques in machine learning, data science, analytics and data mining. They are transparent as they can be presented as rules for human understanding. Explainable AI, human-in-the-loop developments, and fair and ethical AI favour the transparency of decision trees. With the availability of big data, algorithms for constructing decision trees in a platform like MapReduce are a new challenge. This project consists of developing the parallelisation algorithms and their implementation for Hadoop when we consider a forest of decision trees. This implementation will be focused on the applications of the forest of decision trees on privacy-preserving data mining on online social networks (OLSN). We can provide references to recent literature on how the construction of a forest of decision trees is used to provide privacy alerts to users of OLSN. Another application is feature selection. |
|
Ethical Ai Agent The aim is to develop a demonstration of interaction by a human being with an artificial agent about ethical dilemmas and reason over ethical challenges. The agent shall apply several techniques to argue and explain ethical decisions, including game theory or dialogue with humans. Another possibility is deontic logic. The goal is to investigate practical debates between and artificial agents and humans regarding the alignment problem. Possibilities are to explore the use of generative AI and in particular, large language models. |
|
Verifiable AI The goal is to explore mechanisms of formal verification of agents' behaviour, particularly agents who can combine reasoning with operating over a state machine. Moreover, if the behaviour or the knowledge for reasoning is learned, behavioural properties should be verified regarding the behaviour. The aim is to create trustable AI systems because they have been formally verified even if they evolve by learning. The immediate applications are behaviours of autonomous vehicles, and particular aspects could be efficient negotiation of intersections. There is already evidence that artificial agents can negotiate intersections faster than humans and even minimise the requirement of traffic signals. However, there are challenges is mixed environments, such as humans and autonomous vehicles. The impact on mobility of artificial intelligence tools is the focus of this project. |
|
Statistical Modeling of Online Discussions Online discussion is a core feature of numerous social media platforms and has attracted increasing attention for different and relevant reasons, e.g., the resolution of problems in collaborative editing, question answering and e-learning platforms, the response of online communities to news events, online political and civic participation, etc. This project aims to address, from a probabilistic modeling perspective, some existing challenges, both computational and social, that appear in platforms that enable online discussion. For example, how to deal with scalability issues, how to evaluate and improve the quality of online discussions, or how to generally improve social interaction such platforms. Generative models of online discussion threads: state of the art and research challenges |
|
Abstraction and Reasoning Challenge This project is about addressing the Abstraction and Reasoning Challenge (ARC), a kaggle competition consisting of large corpus of visual tasks that require abstraction and reasoning skills to be solved. Each ARC task contains 3-5 pairs of train inputs and outputs, and a test input for which you need to predict the corresponding output with the pattern learned from the train examples. The objective of this project to apply AI techniques (and possibly suggest novel ones) that combine search and learning for automated program synthesis that are developed in the AI/ML research group to address the challenge, or a part of it. |
|
Solving matching and alignment tasks with Neural Networks In this project you will be working in the project Matching Learning (red.es BMATUPF) in close collaboration with BMAT. BMAT is a music company that uses artificial intelligence to index, monitor and report music usage and ownership data across TVs, radios, venues and digital platforms worldwide. This offer gives the opportunity to work on a highly new and innovative field: the application of state of the art Natural Language Processing tools based on Deep Learning and Large Language Models (GPTx and open source versions of ChatGPT) to process (semi and unstructured) data.. The challenges that you will be addressing will be realistic and relevant for industry as they will be provided directly from BMAT. The research will be centered on the power of mixing Transformers [1,2] and pretrained transformers (like ChatGPT), Graph Neural Networks and Knowledge Graphs for having higher capabilities when solving matching and entity alignment tasks [3,4]. A good level of Python and some experience with Neural Network libraries (keras, tensorflow or pytorch) would be needed. Python notebook will be used. [1] Mikolov et al. Efficient Estimation of Word Representations in Vector Space, 2013 [2] M Leone et al. Attention is all you need [3] N Fanourakis et al. A Critical Re-evaluation of Neural Methods for Entity Align |
|
Graph Neural Networks for Large-scale Optimization Large-scale constrained optimization problems are frequently encountered in real-world applications, e.g. traffic management, logistic planning, warehouse optimization, multiagent planning, and scheduling, to name a few. Though constrained optimization problems have been extensively studied, because of their computational complexity, existing optimal solvers do not scale well to complex problems. Recently, graph neural networks (GNNs) have been successfully applied to approximate solutions to large optimization problems. The aim of the thesis is to study the potential of graph neural networks for large-scale optimization, and test it in realistic applications. |
|
Linearly-solvable Markov decision processes with function approximation A linearly-solvable Markov decision process (LMDP) is a special case of the more commonly used Markov decision process (MDP) in reinforcement learning. In LMDPs, the Bellman equations governing the value function are linear, which typically makes learning more efficient. In spite of being simpler than MDPs, LMDPs are surprisingly expressive and can be used to represent many reinforcement learning benchmarks. In particular, any MDP with deterministic actions can be converted into an LMDP. However, the large body of work on deep reinforcement learning largely ignores LMDPs as a representation. The goal of this project is to develop function approximation algorithms for LMDPs, e.g. deep neural networks. One way to do so is to study existing algorithms for MDPs and simplify them for the LMDP setting.. |
|
Entropy-regularized reinforcement learning algorithms Entropy regularization is an extremely popular component of modern reinforcement learning methods. After several years of applying entropy regularization as a heuristic to drive exploration and improve optimization properties of RL algorithms, there has been lots of recent progress in providing solid theoretical foundations justifying this technique. Most recently, a sophisticated regularization technique has lead to the development of a principled RL algorithm, Q-REPS, that is entirely derived from first principles, yet can be implemented in large-scale environments essentially without any approximations to the theory. This project aims to progress the research on this method from two parallel directions: 1) Implement Q-REPS in large-scale environments and analyze its performance empirically and 2) analyze various properties of the algorithm that remain poorly understood. (Some important questions being: understanding the role of the regularization functions and finding alternatives, or studying the properties of the action-value functions returned by the algorithm.) The project requires strong mathematical skills, particularly in linear algebra and probability theory. Knowledge of convex analysis and optimization is a plus, but not absolutely necessary at the current stage. A unified view of entropy-regularized Markov decision processes |
|
Error averaging in approximate dynamic programming Error-propagation analysis of approximate dynamic programming methods is a classic research topic with several implications on reinforcement learning algorithms. The most well-known results consider basic DP algorithms like value iteration and policy iteration, with several new results added in recent years regarding regularized counterparts of these methods. One discovery on this front was the discovery that certain regularization choices result in an "error-averaging" property that ensures a benign propagation of policy evaluation errors: roughly speaking, instead of accumulating the absolute errors linearly, these schemes allow the cancellation of positive and negative errors. This project sets out to understand this phenomenon in more detail by decoupling the averaging effect from the regularization effects as much as possible, thus potentially enabling the development and analysis of more effective error-averaging RL algorithms. The project requires strong mathematical skills, particularly in linear algebra and probability theory. Knowledge of convex analysis and optimization is a plus, but not absolutely necessary at the current stage. Approximate Modified Policy Iteration and its Application to the Game of Tetris Leverage the Average: an Analysis of KL Regularization in RL |
|
Thompson sampling for sequential prediction Thompson sampling is one of the most well-studied algorithms for a class of sequential decision-making problems known as stochastic multi-armed bandit problems. In a stochastic multi-armed bandit problem, a learner selects actions in a sequential fashion, and receives a sequence of rewards corresponding to the chosen actions. A crucial assumption made in this problem is that the rewards associated with each action are random variables drawn independently from a fixed (but unknown) distribution. The goal of this project is to do away with this assumption and study Thompson sampling in non-stationary environments where the rewards may be generated by an arbitrary external process. Precisely, the project considers the framework of sequential prediction with expert advice, and aims to show theoretical performance guarantees for this algorithm and/or analyze its perormance empirically. The project requires strong mathematical skills, particularly in probability theory and multivariate calculus. Knowledge of convex analysis is a plus, but not absolutely necessary at the current stage. Learning to Optimize Via Posterior Sampling An Information-Theoretic Analysis of Thompson Sampling |
|
Better algorithms for online linear-quadratic control Linear-quadratic control is one of the most well-studied problem settings in optimal control theory: it considers control systems where states follow a linear dynamics, and the incurred costs are quadratic in the states and control inputs. In recent years, the problem of online learning in linear-quadratic control problems have received significant attention within the machine-learning community. One particularly interesting development is the formulation of the control problem as a semidefinite program (SDP), which allows the application of tools from online convex optimization. The present project aims to develop new algorithms for online linear quadratic control based on this framework by exploring the possibility of using regularization functions that make better use of the SDP geometry than existing methods based on online gradient descent. The project requires strong mathematical skills, particularly in multivariate calculus and linear algebra. Knowledge of control theory, convex analysis and optimization is a plus, but not absolutely necessary at the current stage. |
|
Deep Neural Multilingual Lexical Simplification Lexical Simplification is a sub-task of Automatic Text Simplification that aims at replacing difficult words with easier to read (or understand) synonyms while preserving the information and meaning of the original text. This is a key task to facilitate reading comprehension to different target readerships such as foreign language learners, native speakers with low literacy levels or people with different reading impairments (e.g. dyslexic individuals). This master’s project aims at investigating novel techniques in Deep Learning and Natural Language Processing to improve current lexical simplification systems. The master student will have the opportunity to develop this research on an available multilingual dataset (English, Portuguese, Spanish) – TSAR 2022 Shared Task Dataset – annotated by human informants. More specifically, the annotations for a given complex word in a sentence is a list of suitable substitutes which could be used instead of the complex word. The candidate is expected to develop cross-lingual or multilingual techniques using available pre-trained models. In the recent TSAR 2022 shared task challenge, several teams participated proposing deep learning techniques (e.g. neural language models) to solve the problem. The candidate is expected to have some knowledge of Natural Language Processing (NLP) and experience with current deep learning paradigms in NLP. ALEXSIS: A Dataset for Lexical Simplification in Spanish. Lexical simplification benchmarks for English, Portuguese, and Spanish. |
|
From Sign Language to Spoken Language: Experiments in Deep Multimodal Translation In the era of mass communication and wide uptake of digital technologies amongst the general public, there still exist numerous communication barriers for the Deaf and Hard-of-Hearing (DHH) community. In spite of the recent advances in Machine Translation (MT) for spoken languages, automatic translation between spoken and Sign Languages or between Sign Languages remains a difficult problem. There is a great opportunity for research in Machine Translation (MT) to bridge the gap between written/spoken languages and Sign Languages. This master’s project aims at experimenting with current approaches in machine translation for sign languages with emphasis on the Sign Language - Spoken Language direction. The student will first experiment with available well-known datasets (e.g. German Sign Language PHOENIX14T dataset – video/text/glosses) exploring different models to extract patterns from videoframes and decode them to produce spoken text. After that, it is planned to apply Transfer Learning to produce models for less resourced SLs (e.g. Spanish or Catalan Sign Language). The candidate is expected to have some knowledge of Natural Language Processing (NLP) and experience with current deep learning paradigms in NLP. Knowledge of image processing is highly desirable. |
|
Audio-visual speech and singing voice separation Source separation is the automatic estimation of the individual isolated sources that make up the audio mixture. The goal of this project is to separate a human voice in a mixture by using both the audio and video modalities. We are interested in both speech and singing voice signals. The most direct applications of speech separation are speaker identification and speech recognition (for example, to create automatic captioning of videos). While some of the applications of singing voice separation are: automatic creation of karaoke, music transcription, or music unmixing and remixing. Leveraging visual and motion information from the target person's face is particularly useful when there are different voices present in the mixture. Deep neural networks that extract features from the video sequence will be explored and used in conjunction with an audio network in order to improve the audio source separation task by incorporating visual cues. |