Research Scientist

Office: 3.B.10

Spemannstr. 41

72076 Tübingen

Germany

Spemannstr. 41

72076 Tübingen

Germany

+49 7071 601 943

+49 7071 601 552

Subject motion during MRI scans can cause severe degradations of the image quality. The problem of correcting for such motion artifacts is one of the most important remaining problems in the field to be solved. Even a few millimeter displacement of the imaged object is enough to generate motion artifacts, which usually appear as ghosts and blur, and make a scan unacceptable for a medical analysis. The acquisition time of high resolution scans can be of an order of minutes, which together with a requirement to keep motionless in millimeter scale is a challenge even for healthy and cooperative subjects. Patients with movement disorders, elder and child population are particularly prone to motion during the image acquisition, while at the same time being the categories of patients who are likely to benefit from MR diagnostics.

Together with my colleagues I am developing data-driven retrospective motion correction algorithms aimed at improving the quality of 3D MRI scans affected by both rigid and non-rigid motion. The crucial aspect of our algorithms is that they require no information about the displacements of a patient in the scanner, i.e. guiding from the tracking cameras. Furthermore, the developed techniques use the raw data from standard imaging sequences requiring no modifications in the scanning pipeline. Importantly, the methods are implemented to run on the graphic cards in order to attain short computation times, which are on the order of seconds when rigid motion is assumed, and on the order of minutes when more complicated non-rigid motion needs to be corrected.

In recent years, most successful motion correction techniques have been based on prospective motion correction. However, prospective methods have certain limitations, which make their use not possible in certain scenarios. The goal of this project is to develop a new retrospective motion correction algorithm that does not require tracking equipment, and can be applied to already acquired data in a post-processing way. Quite importantly, as opposed to prospective techniques the method preserves the original motion corrupted image, which allows to compare the motion correction results with original image.

The current autofocusing-based retrospective methods suffer from the problem of being too slow to be used in the clinical practice. This is because the space of motion parameters is explored in a brute force inefficient way leading to highly suboptimal solutions or long computation times in high-resolution 3D scenarios due to the curse of dimensionality. Together with my colleagues I propose an autofocusing method that is based on a minimization of a cost function that characterizes the quality of the optimized image. The idea is to find the point in the space of possible motions such that the motion corrected image yields the minimum value of a cost function. The challenge for our approach is the fact that for high-resolution 3D volumes the optimization space is vast since there are 6 free parameters per phase encode step. We address this problem by using an analytic model for both motion degradation and its inverse, which allows to find the partial derivatives (gradients) of our cost function, and do an efficient optimization of the objective. For this reason we call our algorithm GradMC (Gradient-based Motion Correction). GradMC starts with an initial estimate of the motion trajectory (which can be no-motion indicating zeros or some small random values), and iteratively corrects for translation and rotation until no more progress in terms of image quality can be made. The image quality metric that we use in our cost function is the entropy of spatial image gradients (first finite derivatives).

We have developed a GPU version of GradMC featuring a parallel implementation of our forward model. For volumes of matrix sizes commonly encountered in clinical practice, only a few minutes are needed for the reconstruction (see Figure 1B). GradMC can correct for arbitrary motion trajectories in 6 degrees of freedom for both 2D and 3D acquisitions. Shown in Figure 1A are the motion correction results on a pair of 2D images acquired with a RARE sequence (matrix size N=384×160).

The Matlab implementation of GradMC along with 4 examples is available at: http://mloss.org/software/view/430/

This project is an extension of the blind retrospective motion correction method aimed at correction of the non-rigid motion. Once again, the proposed method is purely image data-driven meaning that the motion correction algorithm (called GradMC2) does not require information about motion from external sensors as input. Being able to do such a reference-free correction in case of complicated spatially-varying motion is the main contribution we make with current work. The method can be used with generic acquisition sequences, and is associated with a reasonable computation time. We approximate non-rigid motion by locally (patch-wise) rigid motions, and extend the forward model to accommodate this “multi-rigidness” assumption. The method is autofocusing-based, which means that at the core is a generic image quality functional. Given the high complexity of the problem to be solved, and large number of unknowns it is crucial to be able to come up with an efficient optimization algorithm to recover both the underlying sharp image and local displacements characterized by distinct sets of motion parameters specified for each patch. We come up with an annealing-based alternating optimization procedure, where we benefit from the closed form formulation of our forward model in order to do efficient optimization using derivatives of the quality functional with respect to both latent image and motion parameters.

As an ultimate goal we see our method to be applied in routine clinical imaging, where the scans are affected by non-rigid physiological motion (e.g. in abdominal scans). In this study, we make a step forward towards this goal, and in our in vivo experiments we correct for multi-rigid finger motion in a wrist imaging setup (see Figure 2), which can be seen as a reasonable simplification of the realistic non-rigid problem.

Although advances in parallel imaging techniques allow reducing the scan time significantly, it still remains in the order of minutes, increasing the probability of head motion to occur. Motion during the acquisition leads to inconsistent k-space data and thus impedes parallel imaging reconstruction techniques, such as GRAPPA. On the other hand, Fourier domain-based retrospective motion correction techniques typically require a fully acquired k-space as input, and thus necessitate the GRAPPA reconstruction to be performed first. This couples the problems of motion correction and GRAPPA reconstruction, yielding a complex problem to solve.

This project aims at developing of a prototype motion correction algorithm that operates on multi-channel undersampled data acquired with a Cartesian sequence which was augmented by a free induction decay navigator motion monitoring module. The algorithm is based on iterating the application of the GRAPPA reconstruction, motion estimation and motion correction. The acquired FIDnavs which were recently shown to be beneficial for retrospective motion correction were used to aid the motion correction process. The preliminary motion correction results are shown in Figure 3.

**Max Planck Institute for Intelligent Systems** - Tübingen, Germany

- 2011–present
- Graduate student

**Max Planck Institute for Biological Cybernetics** - Tübingen, Germany

- 2009–2011
- Graduate student

**Universität Osnabrück**

- 2007–2009
- M. Sc. in Cognitive Science
- Focus: Neuroinformatics/Robotics

**Volgograd State Pedagogical University** - Volgograd, Russia

- 2001–2005
- B. Sc. in Physics
- Focus: Development of Physics educational software

12 results
(BibTeX)

**Automatic detection of motion artifacts in MR images using CNNS**
*IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2017)*, pages: 811-815, 2017 (conference)

**Improving Quantitative Susceptibility and R2* Mapping by Applying Retrospective Motion Correction**
*23rd Annual Meeting and Exhibition of the International Society for Magnetic Resonance in Medicine*, ISMRM, June 2015 (poster)

**Retrospective rigid motion correction of undersampled MRI data**
*23rd Annual Meeting and Exhibition of the International Society for Magnetic Resonance in Medicine*, ISMRM, June 2015 (poster)

**Retrospective motion correction of magnitude-input MR images**
*International Conference on Machine Learning (ICML) 2015, Workshop on Machine Learning meets Medical Imaging*, 9487, pages: 3-12, Lecture Notes in Computer Science, (Editors: K. K. Bhatia and H. Lombaert), Springer, First International Workshop, MLMMI, July 2015 (conference)

**Blind Retrospective Motion Correction of MR Images**
University of Tübingen, Germany, May 2015 (phdthesis)

**Blind multirigid retrospective motion correction of MR images**
*Magnetic Resonance in Medicine*, 73(4):1457-1468, April 2015 (article)

**FID-guided retrospective motion correction based on autofocusing**
Joint Annual Meeting ISMRM-ESMRMB, Milano, Italy, 2014 (poster)

**Blind Retrospective Motion Correction of MR Images**
*Magnetic Resonance in Medicine (MRM)*, 70(6):1608–1618, 2013 (article)

**Existential neuroscience: a functional magnetic resonance imaging investigation of neural responses to reminders of one’s mortality**
*Social Cognitive and Affective Neuroscience*, 7(2):193-198, 2012 (article)

**Blind Retrospective Motion Correction of MR Images**
20th Annual Scientific Meeting ISMRM, May 2012 (poster)

**Retrospective blind motion correction of MR images**
*Magnetic Resonance Materials in Physics, Biology and Medicine*, 24(Supplement 1):498, 28th Annual Scientific Meeting ESMRMB, October 2011 (poster)

**Automatic foreground-background refocusing**
In pages: 3445-3448, (Editors: Macq, B. , P. Schelkens), IEEE, Piscataway, NJ, USA, 18th IEEE International Conference on Image Processing (ICIP), September 2011 (inproceedings)