Skip to main content
  • Original research
  • Open access
  • Published:

Evaluating different methods of MR-based motion correction in simultaneous PET/MR using a head phantom moved by a robotic system



Due to comparatively long measurement times in simultaneous positron emission tomography and magnetic resonance (PET/MR) imaging, patient movement during the measurement can be challenging. This leads to artifacts which have a negative impact on the visual assessment and quantitative validity of the image data and, in the worst case, can lead to misinterpretations. Simultaneous PET/MR systems allow the MR-based registration of movements and enable correction of the PET data. To assess the effectiveness of motion correction methods, it is necessary to carry out measurements on phantoms that are moved in a reproducible way. This study explores the possibility of using such a phantom-based setup to evaluate motion correction strategies in PET/MR of the human head.


An MR-compatible robotic system was used to generate rigid movements of a head-like phantom. Different tools, either from the manufacturer or open-source software, were used to estimate and correct for motion based on the PET data itself (SIRF with SPM and NiftyReg) and MR data acquired simultaneously (e.g. MCLFIRT, BrainCompass). Different motion estimates were compared using data acquired during robot-induced motion. The effectiveness of motion correction of PET data was evaluated by determining the segmented volume of an activity-filled flask inside the phantom. In addition, the segmented volume was used to determine the centre-of-mass and the change in maximum activity concentration.


The results showed a volume increase between 2.7 and 36.3% could be induced by the experimental setup depending on the motion pattern. Both, BrainCompass and MCFLIRT, produced corrected PET images, by reducing the volume increase to 0.7–4.7% (BrainCompass) and to -2.8–0.4% (MCFLIRT). The same was observed for example for the centre-of-mass, where the results show that MCFLIRT (0.2–0.6 mm after motion correction) had a smaller deviation from the reference position than BrainCompass (0.5–1.8 mm) for all displacements.


The experimental setup is suitable for the reproducible generation of movement patterns. Using open-source software for motion correction is a viable alternative to the vendor-provided motion-correction software.


Simultaneous positron emission tomography and magnetic resonance (PET/MR) systems [1, 2] allow the registration of movements due to the high spatial and temporal resolution and the good contrast of the MR images [3, 4]. This makes it possible to correct the PET data, which is recorded simultaneously, e.g. regarding attenuation correction and patient motion [5,6,7]. There are several approaches regarding the correction of head movements during PET acquisition [8,9,10,11,12,13,14,15,16,17,18].

Due to the comparatively long measuring time in PET, patient movements are inevitable [19] and lead to artifacts, so-called blurring, which influence the imaging process. Movements, which can be divided into two types, further worsen the already comparatively low resolution of PET. In addition to strong, rapid movements of the head, there are also slow drift movements, which result, for example, from the relaxation of the patient’s muscles in the device [20]. Changes in the position of organs are also accompanied by movements of the lesions. This leads to misinterpretations of the tracer uptake, lesion size, decreased PET image quality and quantitative accuracy [21, 22]. Hence, applying some form of motion correction (MoCo) is recommended in PET studies [23]. The mentioned blurring effect leads to increase in the volume in lesions or in the case of this work of a flask in a phantom and serves as a measure of the quality of the MoCo.

Different methods for MoCo are possible, such as vendor-provided proprietary algorithms and custom offline solutions based on open-source software. The methods presented here use the echo-planar imaging (EPI) MR sequence for MoCo [21], which offers the possibility of registering movements with a high temporal resolution based on a rigid registration [24]. Alternatively, PET-only methods use non-attenuation corrected (NAC) or attenuation corrected (AC) images to estimate the motion [25,26,27]. Current methods for correcting movements in PET/MR show an improvement in PET image quality [4]. However, an evaluation of these procedures is difficult in a clinical setting using patient data. To evaluate such procedures, the registered movement in the PET/MR system must be known a priori and reproducible. This condition is not feasible for patient movement.

This study describes a phantom-based experimental setup for evaluation of novel procedures for MoCo. We have divided the paper into two parts, firstly the evaluation of two MoCo methods and a specially created motion-adapted algorithm for framing of the data, and secondly several algorithms were compared with regard to their tracking properties of the phantom movement.


Experimental setup

An MR-compatible robotic system (INNOMOTION, Innomedic GmbH, Herxheim, Germany, Fig. 1), which was initially developed to guide interventions, was used to generate rigid motion of a head phantom. The robot arm is pneumatically driven in 5 DOF (Degrees Of Freedom) [28]. Attached to a 180\(^\circ\) ring, which can be fixed to the patient table, the arm can be manually prepositioned into the orbit region, at fixed angles [29]. Before taking a series of measurements, software-controlled initialisation and referencing was always carried out. The robotic arm has a maximum range of 150 mm in axial direction and a transverse range of 300 mm at holder (restricted by the magnet bore) and, according to manual, an accuracy of ± 1 mm and ± 1\(^\circ\) when positioning.

Fig. 1
figure 1

Schematic representation of the InnoMotion robotic system. The possible directions of movement are indicated by the coloured arrows. Translational movements were performed along the z-axis and rotations around the x-axis (\(\theta _1\)). Innomedic GmbH [28]

Fig. 2
figure 2

Commercial skull model and head phantom (left, [30]), corresponding \(T_1\) MPRAGE of the head phantom (middle, for the sequence parameters see Table 1) and PET image of the filled flask within the phantom (right). The PET image (right) shows one of the three markers

A realistic head phantom [30] with an embedded lesion (fillable flask, 50 ml) was used (Fig. 2). This phantom was modeled to have a structure corresponding to the anatomy of the human skull, as well as attenuation and MR properties similar to human tissue. The phantom is equipped with a flask with a volume of 50 ± 0.12 ml at 20 \(^\circ\)C (VITLAB ©), which can be closed and is accessible from the outside. The flask is located in the brain region. For more information on this phantom (materials, construction and experimental evaluation), please refer to the relevant publication [30]. The flask was used as the basis for evaluating the quality of the MoCo processes. Since lesion movements result, among other consequences, in an increase in the resulting lesion volume, this was also used as an evaluation standard. The movement of the robotic arm could be transferred to the head phantom by a custom-made device consisting of plastic elements (Fig. 3).

Fig. 3
figure 3

InnoMotion robot system and experimental setup for generating a translation motion (left). Extension of the setup to create rotations around the transverse axis (right). The pulling motion of the robot arm can be converted into a rotational motion (see red arrows)

It was possible to move the robot arm in z-direction (longitudinal axis) via the vendor-provided software (MotionCheck, V. By extending the setup (Fig. 3), an additional rotational movement around the x-axis (transversal axis, head tilts towards or away from the chest; \(\theta _1\) in Fig. 1) could be generated.

The movement was tracked using three markers (1.5 ml volume) mounted to the rigid experimental setup. The markers could be filled with a radioactive liquid for subsequent verification of the motion amplitude. By attaching them around the phantom, the same displacement affected the markers as the head phantom and they were visible in the 3D PET volume at the same time.


Before starting a series of measurements, the flask within the phantom as well as the markers were filled with aqueous \(^{18}\)F-FDG (2-Fluoro-2-Deoxy-D-Glucose) solution. The activity was between 30 and 40 MBq at the time of the start of a measurement series. A measurement series (same motion amplitude) consisted of one static image without motion (reference measurement, reference volume), followed by four images taken under the same motion amplitude. This allowed the averaging over these four to estimate the stochastic error.

All measurements were performed using a Siemens Biograph mMR PET/MR (Siemens Healthineers, Erlangen, Germany). For the acquisition of the necessary data, an established clinical protocol was used (Table 1).

Table 1 Listing of the most important sequence parameters

The PET data were aquired in listmode (LM) format. The PET acquisition and EPI sequence ran simultaneously for 600 s each for all data sets. The following settings were used for the PET reconstruction: OSEM (8 iterations, 21 subsets), the Brain HiRes as \(\mu\)-Map, a relative scattering correction, a 256  \(\times\)  256 image matrix, 2.8 zoom and a Gauss-Filter with 3 mm FWHM. This resulted in a voxel size of 1.0 \(\times\) 1.0 \(\times\) 2.03 mm\(^3\) after reconstruction.

Data-processing methods

The following MoCo algorithms were evaluated: BrainCompass, which is part of the Biograph mMR, and MCFLIRT (FMRIB Software Library [31,32,33]). Datasets were reconstructed on the console into multiple frames without MoCo and into a single frame using BrainCompass. MCFLIRT was incorporated into the clinical routine via a framework (Fig. 4) to estimate the motion from the EPI images.

Fig. 4
figure 4

Framework for reconstruction of equidistant frames using MCFLIRT. (1) The PET data was split into frames and reconstructed by the manufacturer software (Siemens). (2) The MR data (EPI) were spatially and temporally adapted to the PET data. EPI data were interpolated by miconv (ODIN [35]) which uses an Akima spline [40] for this purpose. (3) MCFLIRT was executed based on the adapted EPI data. (4) The resulting transformation matrices are then applied to the PET frames. (5) Finally, all motion-corrected frames are summed

Fig. 5
figure 5

Framework for motion-adapted reconstruction using MCFLIRT. (1) MCFLIRT was applied to the EPI data. (2) An algorithm then determined the frames to be reconstructed. (3) Based on the defined movement intervals, the PET data were reconstructed on the Siemens console. (4) The framework from Fig. 4 was used to obtain a motion-corrected image

For MCFLIRT, different schemes for dividing the LM data into time frames were investigated. As a simple strategy, the PET data were divided into equidistant frames (30 \(\times\) 20s frames). Alternatively, it is also possible to divide the PET measurement data into frames based on the motion that occurred [20]. The framework shown in Fig. 4 has been extended for this purpose (Fig. 5). The motion-adapted frame-splitting method detects sudden movements with an amplitude of more than 0.5 mm and uses the registered motion jumps as frame boundaries. The comparison is made using successive averaged translation values:

$$\begin{aligned} |d_{n}-d_{n-1}| \ge 0.5\, \text {mm} \end{aligned}$$

with \(d_{n}\) is the absolute translation displacement calculated by MCFLIRT for the n-th EPI frame (2 s sampling). Frames of less than 20 s were appended to the previous frame (1\(^{st}\) frame excluded) to avoid frames which are too short. To correct slow drift movements, a frame is prevented from lasting longer than 2 min by splitting the interval in the middle.

The default settings for all MCFLIRT-based results were trilinear interpolation and normalized correlation as cost function. These settings have not been changed for MoCo since they were found to provide best results in preliminary tests. To avoid any interference due to differences in PET reconstruction, the static volume (1st image of a measurement series) served as a reference for the volumes determined by MCFLIRT (motion-adapted frames and 30  \(\times\)  20s frames).

For the comparison of the registered motion, the motion parameters had to be extracted. MCFLIRT presents the motion parameters both in a (4 \(\times\) 4)-matrix (.mat file) and in a list format (.par file) of the form rot\(_{\alpha }\), rot\(_{\beta }\), rot\(_{\gamma }\), trans\(_x\), trans\(_y\), trans\(_z\) per frame. The .par file was used for this work.

The BrainCompass, which uses the EPI sequence for motion correction, stores the motion parameters in the DICOM header in a separately created folder (MoCoSeries) for each dicom file ((0020,4000) LT [Motion: trans\(_x\), trans\(_y\), trans\(_z\), rot\(_{\alpha }\), rot\(_{\beta }\), rot\(_{\gamma }\)]). The exact registration algorithm is not published in detail, but some information can be found in [34]. According to this, PET data is divided into individual motion states corresponding to the patient’s movement. The \(\mu\)-Map is adapted to the motion frames and after the reconstruction of all frames, a transformation to a reference PET frame takes place with subsequent summation.

Movement patterns

Three types of motion patterns were generated for the evaluation of the different algorithms. Firstly, simple translational motions were generated along the z-axis (see Fig. 1) with different maximum motion amplitudes ranging from 0 to 20.24 mm. The single motion was performed 300 s after the start of acquisition, i.e., in the middle of the PET acquisition. Secondly, rotational movements around the x-axis (\(\theta _1\) in Fig. 1), with rotation angles between 0\(^\circ\) and 3.49\(^\circ\), were generated. Again, the motion was initiated at 300 s. Finally, a more complex motion pattern compared to a simple translational motion was generated to verify motion-adapted frame-splitting (for illustration see Fig. 10) prior to MCFLIRT. The goal was to reach the final position by stepwise translational displacement along the z-axis. Eight steps of 0.5 mm translational displacement were performed consecutively, starting 180 s after the beginning of the PET acquisition. There were 20 s between each of the individual displacements. Alternative 1 and 2 were generated as a result of the described method. In Alternative 1, the frame boundaries were set after each detected motion and the areas before and after the gradual movement (motion pattern) remained without further subdivision. Alternative 2 allows a finer subdivision of the reconstructed frames even when no motion was detected. These different subdivisions were achieved by adjusting the motion-adapted algorithm by changing or removing the maximum time span of 2 min. A slow drift movement was not to be expected with the described phantom setup.

Evaluation method

For the evaluation of the datasets and the determination of the motion-corrupted and corrected volumes (comparative value), evaluation programs were written in Python and UNIX shell scripts. Furthermore, the Object-oriented Development Interface for NMR (ODIN version 2.0.4, [35]) was used to transform image data and perform arithmetic calculations with whole datasets. Four datasets were available for evaluation: the motion-uncorrected PET image, the motion-corrected PET images from BrainCompass and MCFLIRT with equidistant frames and, if relevant, motion-adapted frame splitting (Fig. 6). The following is a brief description of the procedure of the program. First, the markers had to be removed from the PET image using spherical masks. The maximum activity concentration \(A_\text {max}\) and a lower threshold \(0.1 \cdot A_\text {max}\) was then determined. Using this threshold, all voxels in the interval \([0.1 \cdot A_\text {max}, A_\text {max}]\) were summed up and the result characterized the volume. The lower threshold was selected based on preparatory evaluations which showed that the calculated volume was relatively insensitive to changes in the specific value of the threshold below 0.17. In a final step, all volumes were normalized to the reference volume (first measured value of a measurement series, not corrupted by movement). To check the plausibility of the movement generated by the robotic arm, it was possible to check the positions of the individual markers by a COM algorithm before and after the motion.

Fig. 6
figure 6

Schematic representation of the evaluation method. After reconstruction, three or four datasets were available: motion-uncorrected, BrainCompass and MCFLIRT (30 \(\times\) 20s and if relevant motion-adapted). To obtain MCFLIRT corrected images, the frameworks of Figs. 4 and 5 had to be applied

In the case of the motion-adapted reconstruction, the calculation time and memory requirements were compared with the MCFLIRT standard method (30 \(\times\) 20s frames). To determine the calculation time, time stamps were incorporated into the framework and finally a difference was formed. The calculation time refers only to the registration and does not consider the reconstruction of the individual frames.

In addition to looking at volumes, the maximum activity concentration (\(A_{max}\)) and center-of-mass (COM) were used to evaluate the MoCo processes. The COM calculated on the basis of the PET image was determined before and after the motion, as well as after the application of BrainCompass and MCFLIRT. Similarly, \(A_{max}\) was determined. ODIN was used to determine the COM and \(A_{max}\) in each case. For the observation of \(A_{max}\) it was necessary to perform a decay correction which was realized by a shell script using ODIN for each image.

Comparison of different registration methods

Due to the well-known phantom motion, other registration methods could be considered in a comparison regarding motion tracking (see Figs. 11 and 12). Different motion estimates were compared with the robot-induced motion (ground truth). The open-source Synergistic Image Reconstruction Framework (SIRF) [36] was used to reconstruct PET images without attenuation correction (NAC) and estimate motion using SPM12 [37] and NiftyReg [38, 39]. Compared to MCFLIRT, the PET Raw Data in LM format were used here, so the workflow is independent of the Siemens software. During the reconstruction of the NAC images, the data were divided into the corresponding frames (30 \(\times\) 20s or motion-adapted). The OSEM algorithm with 8 iterations and 21 subsets was used, taking into account randoms, detector sensitivity and scattering. This resulted in a voxel resolution of 2.32 \(\times\) 2.32 \(\times\) 2.03 mm\(^3\) after reconstruction.

Furthermore, all algorithms (MCFLIRT, BrainCompass, SPM12 and NiftyReg) were compared with respect to registration based on EPI. The registrations based on EPI data have a temporal resolution of 2 s, while NAC-based registrations has used 30 \(\times\) 20s frames. For comparability, the 1st frame was also defined as the reference frame here.

For comparison with MCFLIRT and BrainCompass, the motion parameters were saved in separate files. SIRF offers appropriate methods for saving the parameters in a (4 \(\times\) 4)-matrix. From this, the rotation angles were calculated and the translation values extracted.


Translation motion

Figure 7 shows results for an acquisition with several translations of different amplitudes in the longitudinal direction. The specified movement parameters all originate from MCFLIRT and have therefore been detected in the same coordinate system. Each manufacturer, each system and the various algorithms usually use different coordinate system origins (for PET and MR). This does not affect the absolute translation vector, but it does affect the registered rotation angles. To avoid errors here, the translation shift determined by MCFLIRT was always given as a “standard” measure. Thus, the results are comparable. The error with respect to the x-axis resulted from the standard deviation (SD) which was calculated by averaging over the absolute translation displacement of the measurement series. This can be caused by deviations in the execution of movements by the robotic arm or by deviations in the motion registration. The error bars that corresponds to the volumes are the SD with regard to the volume determination results.

Fig. 7
figure 7

Results of MoCo procedures for different translation amplitudes by pulling the head phantom. Shown are the motion-uncorrected volume (black), BrainCompass (green) and MCFLIRT with equidistant frames (blue). The values were averaged over four identical motion amplitudes with the error bar in y-direction as the SD. The error bar in x-direction shows the fluctuations of the measured amplitude (SD)

For the motion-uncorrected case, the estimated volume of the reconstructed lesion (flask) increased as the motion amplitude increased. In contrast, BrainCompass and MCFLIRT corrected the movement. While BrainCompass volumes deviate from the reference volume by approximately 0.7–4.7%, the results of MCFLIRT are between  -2.8–0.4%. The horizontal red line by 100% in Fig. 7 corresponds to an ideal correction and that means a correction in which the reference volume would be reached.

In order to evaluate the results using further assessment metrics, COM (Fig. 8 and the Additional file  1; Supplementary Table 1, 2) and \(A_{max}\) (Fig. 9 and the Additional file  1; Supplementary Table 3) were also considered. Looking at the results for COM (Fig. 8), the distance (euclidean norm) from COM after applying BrainCompass and MCFLIRT to COM before the motion (reference position) is shown for the different set translation amplitudes. The results show that MCFLIRT (ranging from 0.2 to 0.6 mm after MoCo) had a smaller deviation from the reference position than BrainCompass (between 0.5 to 1.8 mm) for all examined displacements. Looking at the results for \(A_{max}\) in Fig. 9 shows the relative deviation from the value before motion for motion-uncorrected, BrainCompass and MCFLIRT. The deviation of MCFLIRT from the reference value (\(A_{max}\) before motion) is less than 1.2% for all translation amplitudes while the BrainCompass provides deviations between 1.3% and 2.2%.

Fig. 8
figure 8

Representation of the absolute deviation of the COM without any MoCo and after MoCo using BrainCompass and MCFLIRT. The red line symbolizes 0 mm distance to the COM before the motion (reference). The values represent the Euclidean distance between two points in space

Fig. 9
figure 9

Representation of the relative deviation of \(A_{max}\) without any MoCo and after application of BrainCompass and MCFLIRT. The red line symbolizes \(A_{max}\) for the reference (before motion)

Fig. 10
figure 10

Schematic representation of the motion-adapted frames for Alternative 1 (left) and for Alternative 2 (right). The blue line represents the absolute translation amplitude and the red lines symbolize the frame boundaries

Rotational motion

Results of BrainCompass and MCFLIRT for rotational movements are listed in Table 2. As expected, the motion-uncorrected volume increased with larger rotation angles, by 1.2–8.9%. However, while the corrected volume after BrainCompass increases with increasing rotation amplitude and for the 1.2\(^\circ\) rotation even exceeds the motion-uncorrected volume, the volumes after MCFLIRT are closer to the reference volume.

Table 2 Volume increase (including SD) of MoCo methods for an increasing angle of rotation compared to the motion-uncorrected volume

Motion-adapted framing

A more complex motion pattern was generated to verify and evaluate the motion-adapted framing using MCFLIRT. Two different alternatives (referred as Alternative 1 and 2, Fig. 10) were created. While the motion pattern led to a relatively high volume increase of about 3.8% (Table 3) in the motion-uncorrected data, BrainCompass delivered an even larger volume of 3.9%. MCFLIRT, with a deviation of 1.2%, was closer to the reference volume. The alternatives showed also a similar deviation of 1.1% (A1ternative 1) and 1.2% (Alternative 2).

Table 3 Volume increase relative to the reference volume (with SD) for the motion pattern in Fig. 10
Table 4 Calculation times and memory requirements for the different alternatives in Fig. 10

Table 4 lists the time and memory required to calculate the motion corrected images with the different subdivisions using MCFLIRT.

Comparison of different registration methods

Fig. 11
figure 11

Comparison of different registration methods on a single translation movement. The robotic movement (black line) symbolizes the ground truth. Upper image: EPI-based; lower image: NAC-based. Shown is the absolute translation displacement depending on time

Fig. 12
figure 12

Comparison of different registration methods of multiple translation movements. The robotic movement (black line) symbolizes the ground truth. Upper image: EPI-based; lower image: NAC-based. Shown is the absolute translation displacement depending on time

Finally, various registration methods were tested based on a single large translation movement (Fig. 11) and on the motion pattern (Fig. 12). The given course of movement of the robotic system served as reference value (ground truth). In the upper plots the different registration methods are compared regarding EPI images, in the lower plots regarding NAC images.

In the upper plot of Fig. 11, all registration methods showed similar results. It shows very small discrepancies between MCFLIRT, BrainCompas, SPM and NiftyReg (all EPI-based). The deviation from the ground truth never exceeds 0.5 mm. In the lower image, the registrations based on the NAC images detected the maximum motion amplitude, but showed a slight fluctuation.

Figure 12 shows that the EPI-based results detected the individual movement steps. The deviations between the ground truth and MCFLIRT or SPM12 remained smaller than 0.5 mm. Only for the motion peaks the difference was larger than 0.5 mm. The BrainCompass could not detect all motion steps and underestimated the maximum displacement by about 1 mm. NiftyReg detected the movement, but overestimated the maximum translation displacement. In the NAC-based methods, it was found that SPM12 overestimated the maximum displacement, while NiftyReg detected a too low amplitude.


Compared to similar recent publications, e.g. [15, 17], a more realistic phantom (with bone analogues and anatomical structures) was used in this study so that motion-correction methods can be evaluated with reduced error in attenuation correction and/or without the need for synthetic \(\mu\)-Maps. Furthermore, several translational and rotational movements were recorded. The use of an MR-compatible robotic system, which was initially intended for interventions, made it possible to generate reproducible real-world (i.e. non-synthetic) movement sequences. The aim of our work was to evaluate different algorithms. For this purpose, the BrainCompass and a MCFLIRT framework, which were integrated into the clinical workflow, were compared. Using SIRF, it was subsequently possible to compare EPI- and NAC-based approaches using SPM12 and NiftyReg with regard to movement registration.

Comparison of BrainCompass and MCFLIRT

When analyzing Fig. 7, Tables 2 and 3 , it becomes apparent that, in the motion-uncorrected case, the volume increases with increasing movement amplitude. This was expected because the activity concentration is smeared over a larger volume due to motion. Both BrainCompass and MCFLIRT corrected the volume, with MCFLIRT providing results closer to 100% of the reference volume. The corrected volumes differ between BrainCompass and MCFLIRT by approximately 1.3-7.4%. At this point, it should be noted that the exact algorithm of the BrainCompass image reconstruction is proprietary, and the cost function, optimization algorithm and interpolation that are employed are generally unknown to the end user.

The inspection of the COM (Fig. 8) shows that both BrainCompass and MCFLIRT corrected the flask to its original position (reference position before motion). While the results of the COM confirm the results from the volume analysis, the results regarding \(A_{max}\) (Fig. 9) show only a small change in \(A_{max}\) after motion. The reason for this was the overlap of the volume before and after the motion. The flask is about 5 cm in diameter at the thickest point, and the volumes overlap in the resulting 3D image when moving up to a maximum of 20 mm. Thus, a MoCo by BrainCompass or MCFLIRT did not provide any improvement with respect to \(A_{max}\). It is striking to look at the results for the adjusted 18 mm displacement, this value is also visible in Fig. 7 (uncorrected). The reason was probably insufficient mixing of FDG and water prior to the measurement series at 18 mm. Thus, in the reconstructed image, there is an increased concentration of activity in the upper part, which leads to a bigger impact due the motion, as the translation amplitude was greater than the extent of activity accumulation. However, the MoCo methods also show an improvement here in the form of the restoration of \(A_{max}\).

To check the motion-adapted frame-splitting algorithm (Fig. 5) for its suitability, a motion pattern (Fig. 10) was designed and applied. The algorithm detected the individual amplitude jumps (movements) and set them as frame boundaries. The volumes (Table 3) based on the motion-adapted subdivision were nearly identical to the standard subdivision (30 \(\times\) 20s frames). Furthermore, the BrainCompass showed no improvement compared to the motion-uncorrected volume.

The duration of the calculations and the memory requirements were also compared (Table 4). As expected, with a lower number of frames, the calculation time as well as the storage space requirement decreases. This is due to the reduced number of required calculations (time saving) and fewer intermediate results (memory space saving). Furthermore, the occurrence of strong movements within a frame is avoided by motion-adapted frame splitting. This is supposed to prevent the blurring effect from occurring within a frame, which makes the calculation of motion parameters more accurate at least with regard to a NAC-based registration.

The direct comparison of both correction methods shows that both, BrainCompass and MCFLIRT, deliver motion-corrected images. One advantage of MCFLIRT is its open-source license. However, to obtain the results in the presented format, it was necessary to integrate the MCFLIRT into the clinical routine via a framework, in the form of a DICOM node. This is opposed by BrainCompass, an algorithm distributed by Siemens, which is available as part of the PET/MR system, i.e. it is easy to use and certified. The disadvantages are, on the one hand, the necessity of a proprietary license and, on the other hand, the missing information of the exact algorithm and the missing possibility of modifying settings like interpolation or cost function. Both algorithms functioned during all measurements without bugs, failures or similar.

Comparison of different registration methods

In the Figs. 11 and 12 the motion registration hardly differs between MCFLIRT and SPM (EPI-based). Figure 11 shows that the registration using the NAC images (NiftyReg and SPM) is almost identical to the ground truth, with some fluctuations, likely due to the poor SNR on which the registration is based. It should be taken into account that a larger volume (brain image) was available for EPI-based registration, but PET-based could only use the flask and markers for registration.

Due to the better temporal resolution of the EPI data, peaks can be seen in the EPI motion tracking (Figs. 10 and 12), which were the result of movements when the phantom moved. The PET image or the individual frames weren’t affected by these spikes, which can also be seen in the course of the registrations (lower plots in Fig. 12) based on the NAC images. This is because the temporal resolution of the PET is not high enough to register this short movement.

Limitations of the study

There are several limitations regarding this study: The whole setup (robotic system, phantom) represents a simplification of the actual real-world problem of motion correction of a brain measurement. In the context of this work, only simple motion patterns were used to evaluate complex algorithms. However, patient movement is usually more complex, a point that could not be fully reproduced here. For instance, a continuous slow spatial drift, which is often observed in patients, could not be created with the robotic system. The software of the robotic system (MotionCheck, V. does not allow continuous motion. The anatomy in the phantom is also simple, the flask is well defined on the MR and the PET so there is no anatomic and/or physiological background unlike in a real patient. This limitation is expected to impact the accuracy of the NAC registrations, compared to patient data for radiotracers with distributed uptake in the brain.


The experimental setup is suitable for the reproducible generation of movement patterns. This enabled an evaluation of rigid MoCo methods. In addition to the proprietary software BrainCompass, the open-source software MCFLIRT is a suitable alternative. The evaluation showed that MoCo methods lead to a minimization of volume increase through motion, which may result in a better localization of PET data. Furthermore, an algorithm for motion-adapted reconstruction was presented. The program is primarily used for MoCo that can be adapted to the patient’s movements and thus extends MoCo using MCFLIRT.

Registration quality has been demonstrated with both SPM and NiftyReg, as provided by SIRF. Future work could therefore include a comparison of motion-corrected PET reconstructions using SIRF.

Availability of data and materials

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.


  1. Cherry SR, Louie AY, Jacobs RE. The integration of positron emission tomography with magnetic resonance imaging. Proc IEEE. 2008;96(3):416–38.

    Article  CAS  Google Scholar 

  2. Catana C. Principles of simultaneous PET/MR imaging. Magn Reson Imaging Clin. 2017;25(2):231–43.

    Article  Google Scholar 

  3. Gillman A, Smith J, Thomas P, Rose S, Dowson N. Pet motion correction in context of integrated PET/MR: current techniques, limitations, and future projections. Med Phys. 2017;44(12):430–45.

    Article  Google Scholar 

  4. Fürst S, Grimm R, Hong I, Souvatzoglou M, Casey ME, Schwaiger M, Nekolla SG, Ziegler SI. Motion correction strategies for integrated PET/MR. J Nucl Med. 2015;56(2):261–9.

    Article  Google Scholar 

  5. Catana C, Benner T, van der Kouwe A, Byars L, Hamm M, Chonde DB, Michel CJ, El Fakhri G, Schmand M, Sorensen AG. MRI-assisted pet motion correction for neurologic studies in an integrated MR-PET scanner. J Nucl Med. 2011;52(1):154–61.

    Article  Google Scholar 

  6. Chen KT, Izquierdo-Garcia D, Poynton CB, Chonde DB, Catana C. On the accuracy and reproducibility of a novel probabilistic atlas-based generation for calculation of head attenuation maps on integrated PET/MR scanners. Eur J Nucl Med Mol Imaging. 2017;44(3):398–407.

    Article  Google Scholar 

  7. Rahmim A, Rousset O, Zaidi H. Strategies for motion tracking and correction in pet. PET Clin. 2007;2(2):251–66.

    Article  Google Scholar 

  8. Bloomfield PM, Spinks TJ, Reed J, Schnorr L, Westrip AM, Livieratos L, Fulton R, Jones T. The design and implementation of a motion correction scheme for neurological pet. Phys Medi Biol. 2003;48(8):959.

    Article  Google Scholar 

  9. Mukherjee JM, Lindsay C, Mukherjee A, Olivier P, Shao L, King MA, Licho R. Improved frame-based estimation of head motion in pet brain imaging. Med Phys. 2016;43(5):2443–54.

    Article  CAS  Google Scholar 

  10. Noonan P, Howard J, Hallett W, Gunn R. Repurposing the microsoft kinect for windows v2 for external head motion tracking for brain pet. Phys Med Biol. 2015;60(22):8753.

    Article  CAS  Google Scholar 

  11. Jiao J, Searle GE, Schnabel JA, Gunn RN. Impact of image-based motion correction on dopamine d3/d2 receptor occupancy-comparison of groupwise and frame-by-frame registration approaches. EJNMMI Phys. 2015;2(1):1–15.

    Article  Google Scholar 

  12. Olesen OV, Sullivan JM, Mulnix T, Paulsen RR, Hojgaard L, Roed B, Carson RE, Morris ED, Larsen R. List-mode pet motion correction using markerless head tracking: proof-of-concept with scans of human subject. IEEE Trans Med Imaging. 2012;32(2):200–9.

    Article  Google Scholar 

  13. Olesen OV, Paulsen RR, Hojgaard L, Roed B, Larsen R. Motion tracking for medical imaging: a nonvisible structured light tracking approach. IEEE Trans Med Imaging. 2011;31(1):79–87.

    Article  Google Scholar 

  14. Matsubara K, Ibaraki M, Nakamura K, Yamaguchi H, Umetsu A, Kinoshita F, Kinoshita T. Impact of subject head motion on quantitative brain 15 o pet and its correction by image-based registration algorithm. Ann Nucl Med. 2013;27(4):335–45.

    Article  Google Scholar 

  15. Inomata T, Watanuki S, Odagiri H, Nambu T, Karakatsanis NA, Ito H, Watabe H, Tashiro M, Shidahara M. A systematic performance evaluation of head motion correction techniques for 3 commercial pet scanners using a reproducible experimental acquisition protocol. Ann Nucl Med. 2019;33(7):459–70.

    Article  CAS  Google Scholar 

  16. Kyme AZ, Fulton RR. Motion estimation and correction in SPECT, PET and CT. Phys Med Biol. 2021.

  17. Spangler-Bickell MG, Hurley SA, Deller TW, Jansen F, Bettinardi V, Carlson M, Zeineh M, Zaharchuk G, McMillan AB. Optimizing the frame duration for data-driven rigid motion estimation in brain pet imaging. Med Phys. 2021.

  18. Polycarpou I, Soultanidis G, Tsoumpas C. Synergistic motion compensation strategies for positron emission tomography when acquired simultaneously with magnetic resonance imaging. Math Phys Eng Sci Philos Trans R Soc A. 2021.

  19. Herzog H, Tellmann L, Fulton R, Stangier I, Kops ER, Bente K, Boy C, Hurlemann R, Pietrzyk U. Motion artifact reduction on parametric pet images of neuroreceptor binding. J Nucl Med. 2005;46(6):1059–65.

    Google Scholar 

  20. Ullisch MG, Scheins JJ, Weirich C, Kops ER, Celik A, Tellmann L, Stöcker T, Herzog H, Shah NJ. MR-based pet motion correction procedure for simultaneous MR-PET neuroimaging of human brain. PLoS ONE. 2012;7(11):48149.

    Article  Google Scholar 

  21. Catana C, Drzezga A, Heiss W-D, Rosen BR. PET/MRI for neurologic applications. J Nucl Med. 2012;53(12):1916–25.

    Article  Google Scholar 

  22. Grimm R, Fürst S, Souvatzoglou M, Forman C, Hutter J, Dregely I, Ziegler SI, Kiefer B, Hornegger J, Block KT, et al. Self-gated MRI motion modeling for respiratory motion compensation in integrated PET/MRI. Med Image Anal. 2015;19(1):110–20.

    Article  Google Scholar 

  23. Ikari Y, Nishio T, Makishi Y, Miya Y, Ito K, Koeppe RA, Senda M. Head motion evaluation and correction for pet scans with 18F-FDG in the Japanese Alzheimer’s disease neuroimaging initiative (J-ADNI) multi-center study. Ann Nucl Med. 2012;26(7):535–44.

    Article  Google Scholar 

  24. Ashburner J, Friston KJ. Rigid body registration. Statistical parametric mapping: the analysis of functional brain images. 2007;49–62.

  25. Mourik JE, Lubberink M, van Velden FH, Lammertsma AA, Boellaard R. Off-line motion correction methods for multi-frame pet data. Eur J Nucl Med Mol Imaging. 2009;36(12):2002.

    Article  Google Scholar 

  26. Ye H, Wong K-P, Wardak M, Dahlbom M, Kepe V, Barrio JR, Nelson LD, Small GW, Huang S-C. Automated movement correction for dynamic PET/CT images: evaluation with phantom and patient data. PLoS ONE. 2014;9(8):103745.

    Article  Google Scholar 

  27. Schleyer P, Dunn J, Reeves S, Brownings S, Marsden P, Thielemans K. Detecting and estimating head motion in brain pet acquisitions using raw time-of-flight pet data. Phys Med Biol. 2015;60(16):6441.

    Article  CAS  Google Scholar 

  28. Cleary K, Melzer A, Watson V, Kronreif G, Stoianovici D. Interventional robotic systems: applications and technology state-of-the-art. Min Invasive Ther Allied Technol. 2006;15(2):101–13.

    Article  Google Scholar 

  29. Melzer A, Gutmann B, Remmele T, Wolf R, Lukoscheck A, Bock M, Bardenheuer H, Fischer H. Innomotion for percutaneous image-guided interventions. IEEE Eng Med Biol Mag. 2008;27(3):66–73.

    Article  Google Scholar 

  30. Harries J, Jochimsen TH, Scholz T, Schlender T, Barthel H, Sabri O, Sattler B. A realistic phantom of the human head for PET-MRI. Eur J Nucl Med Mol Imaging Phys. 2020;7:52.

    Google Scholar 

  31. Jenkinson M, Smith S. A global optimisation method for robust affine registration of brain images. Med Image Anal. 2001;5(2):143–56.

    Article  CAS  Google Scholar 

  32. Jenkinson M, Bannister P, Brady M, Smith S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. Neuroimage. 2002;17(2):825–41.

    Article  Google Scholar 

  33. Smith SM, Jenkinson M, Woolrich MW, Beckmann CF, Behrens TE, Johansen-Berg H, Bannister PR, De Luca M, Drobnjak I, Flitney DE, et al. Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage. 2004;23:208–19.

    Article  Google Scholar 

  34. Iagaru A, Hope T, Veit-Haibach P. PET/MRI in oncology: current clinical applications. New York: Springer; 2018.

    Book  Google Scholar 

  35. Jochimsen TH, Von Mengershausen M. Odin-object-oriented development interface for NMR. J Magn Reson. 2004;170(1):67–78.

    Article  CAS  Google Scholar 

  36. Ovtchinnikov E, Brown R, Kolbitsch C, Pasca E, da Costa-Luis C, Gillman AG, Thomas BA, Efthimiou N, Mayer J, Wadhwa P, et al. SIRF: synergistic image reconstruction framework. Comput Phys Commun. 2020;249:107087.

    Article  CAS  Google Scholar 

  37. Friston KJ, Ashburner J, Frith CD, Poline J-B, Heather JD, Frackowiak RS. Spatial registration and normalization of images. Hum Brain Mapp. 1995;3(3):165–89.

    Article  Google Scholar 

  38. Modat M, Cash DM, Daga P, Winston GP, Duncan JS, Ourselin S. Global image registration using a symmetric block-matching approach. J Med Imaging. 2014;1(2):024003.

    Article  Google Scholar 

  39. Modat M, Ridgway GR, Taylor ZA, Lehmann M, Barnes J, Hawkes DJ, Fox NC, Ourselin S. Fast free-form deformation using graphics processing units. Comput Methods Programs Biomed. 2010;98(3):278–84.

    Article  Google Scholar 

  40. Akima H. A new method of interpolation and smooth curve fitting based on local procedures. J ACM (JACM). 1970;17(4):589–602.

    Article  Google Scholar 

  41. Martinez-Möller A, Souvatzoglou M, Delso G, Bundschuh RA, Chefd’hotel C, Ziegler SI, Navab N, Schwaiger M, Nekolla SG. Tissue classification as a potential approach for attenuation correction in whole-body PET/MRI: evaluation with PET/CT data. J Nucl Med. 2009;50(4):520–6.

    Article  Google Scholar 

Download references


Open Access funding enabled and organized by Projekt DEAL. We acknowledge support from the “Strahlenschutzseminar in Thüringen e.V.” and Siemens Healthineers AG (Code F2015-04 or C00229543), whose support made the work on the project possible. This work was partly-funded by the Computational Collaborative Project in Synergistic PET-MR Reconstruction, CCP PET-MR, UK EPSRC grant EP/M022587/1 and its associated Software Flagship project EPSRC grant EP/P022200/1. The integration of MR Robotic System was supported by the BMBF Grant SONO-RAY FKZ: 03Z1L511.

Author information

Authors and Affiliations



EE created the setup, performed measurements, evaluated the data, drafted the manuscript. TJ assisted creating the setup, developed the framework in Fig. 4. JH created the head phantom. AM and MU made it possible to use the robotic system for the purpose of this study. KT and RB assisted in SIRF. BS and OS initiated and designed the study. EE, TJ, OS and BS contributed equally to this work. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Eric Einspänner.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

O.S. serves as a consultant and advisor for Life Molecular Imaging Healthcare, Positrigo AG, TEVA Pharmaceuticals and Drägerwerk. O.S. is a principal investigator for Life Molecular Imaging Healthcare and receives project funding herein. He served as PI for TEVA Pharmaceuticals and Dräger as well as received project funding from these companies.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1. Supplementary Table 1.

The table shows the position (z,y,x) of the COM before and after the motion. The 1st column shows the originally set movement amplitude along the z-axis. Supplementary Table 2. The table shows the position (z,y,x) of the COM after application of BrainCompass and MCFLIRT. The 1st column shows the originally set movement amplitude along the z-axis. Supplementary Table 3. \(A_{max}\) in [Bq/ml] reference (before motion), without any MoCo (motion-uncorrected) and after application of BrainCompass and MCFLIRT. The 1st column shows the originally set movement amplitude along the z-axis.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Einspänner, E., Jochimsen, T.H., Harries, J. et al. Evaluating different methods of MR-based motion correction in simultaneous PET/MR using a head phantom moved by a robotic system. EJNMMI Phys 9, 15 (2022).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: