Hoppa till huvudinnehållet
Till KTH:s startsida

PhD position: AI guided multimodal/multichannel spatiotemporal imaging

The PhD position is on developing mathematical theory and algorithms for foundational models for image reconstruction in multichannel dynamic tomographic imaging. Specific emphasis will be on regularisation for reconstruction of a dynamic image from time series data obtained from dynamic PET/CT and dynamic photon-counting X-ray CT. 

 

Background & motivation

Tomographic imaging techniques, like X-ray computed tomography (CT) and positron emission tomography (PET), probe the object being imaged with a wave/particle from multiple known directions. The data represents indirect noisy observations of the internal structure (=image), so one needs to use techniques from inverse modelling (reverse engineering) to computationally recover the image from the data. Such reverse engineering is often ill-posed, meaning that methods that try to recover the image by fitting observed data only are unstable (small changes in data leads to large changes to the image). Hence, a key element is to stabilise the recovery by use an appropriate regularisation method. Finding an appropriate regularisation is based on using a priori information, and this becomes more challenging when data and images are multidimensional and dynamic. 

 

Research topics

A particular research topic is to set-up, train and evaluate AI based approaches that utilise domain specific physics informed mutlimodal neural network architectures for image reconstruction. The idea is to efficiently integrate complementary information from the different channels in the data and image by coupling them to each other by specific physics informed deep generative models. Additionally, we can view dynamics as an additional (temporal) channel that is modelled with another physics informed deep neural network. A starting point is to study transformer architectures for multichannel imaging based on cross attention and architectures based on state space model architectures (SSM), like Mamba, that predict an output sequence from a continuous input sequence through learned state and output equations. One can here also account for mutliscale phenomena by considering tree-based hierarchical transformers. Furthermore, one can model dynamics as a sequence of deformations, each modelled with a foundational models for image registration. See the research program for more information.

 

Cross-Disciplinary Research Environment

The PhD position is part of a larger collaboration involving Mats Persson at the Physics of Medical Imaging division within the Department of Physics at KTH and Massimiliano Colarieti Tosti at the Division of Biomedical Engineering within the Department of Biomedical Engineering and Health Systems at KTH. They will also recruit one PhD student each that will be working in the project. This is especially the case when focus shifts towards scaling up the AI models so that they can be applied to dynamic PET/CT and spectral CT.

 

Link to apply

Use the link below to apply for this position, deadline is May 5, 2025.

Link: https://www.kth.se/lediga-jobb/804218?l=en


Profilbild av Ozan Öktem

Portfolio

  • PhD position: AI guided multimodal/multichannel spatiotemporal imaging