PURPOSE: Data-driven rigid motion estimation for PET brain imaging is usually performed using data frames sampled at low temporal resolution to reduce the overall computation time and to provide adequate signal-to-noise ratio in the frames. In recent work it has been demonstrated that list-mode reconstructions of ultrashort frames are sufficient for motion estimation and can be performed very quickly. In this work we take the approach of using image-based registration of reconstructions of very short frames for data-driven motion estimation, and optimize a number of reconstruction and registration parameters (frame duration, MLEM iterations, image pixel size, post-smoothing filter, reference image creation, and registration metric) to ensure accurate registrations while maximizing temporal resolution and minimizing total computation time. METHODS: Data from 18 F-fluorodeoxyglucose (FDG) and 18 F-florbetaben (FBB) tracer studies with varying count rates are analyzed, for PET/MR and PET/CT scanners. For framed reconstructions using various parameter combinations interframe motion is simulated and image-based registrations are performed to estimate that motion. RESULTS: For FDG and FBB tracers using 4 × 105 true and scattered coincidence events per frame ensures that 95% of the registrations will be accurate to within 1 mm of the ground truth. This corresponds to a frame duration of 0.5-1 sec for typical clinical PET activity levels. Using four MLEM iterations with no subsets, a transaxial pixel size of 4 mm, a post-smoothing filter with 4-6 mm full width at half maximum, and averaging two or more frames to create the reference image provides an optimal set of parameters to produce accurate registrations while keeping the reconstruction and processing time low. CONCLUSIONS: It is shown that very short frames (≤1 sec) can be used to provide accurate and quick data-driven rigid motion estimates for use in an event-by-event motion corrected reconstruction.
PURPOSE: Data-driven rigid motion estimation for PET brain imaging is usually performed using data frames sampled at low temporal resolution to reduce the overall computation time and to provide adequate signal-to-noise ratio in the frames. In recent work it has been demonstrated that list-mode reconstructions of ultrashort frames are sufficient for motion estimation and can be performed very quickly. In this work we take the approach of using image-based registration of reconstructions of very short frames for data-driven motion estimation, and optimize a number of reconstruction and registration parameters (frame duration, MLEM iterations, image pixel size, post-smoothing filter, reference image creation, and registration metric) to ensure accurate registrations while maximizing temporal resolution and minimizing total computation time. METHODS: Data from 18 F-fluorodeoxyglucose (FDG) and 18 F-florbetaben (FBB) tracer studies with varying count rates are analyzed, for PET/MR and PET/CT scanners. For framed reconstructions using various parameter combinations interframe motion is simulated and image-based registrations are performed to estimate that motion. RESULTS: For FDG and FBB tracers using 4 × 105 true and scattered coincidence events per frame ensures that 95% of the registrations will be accurate to within 1 mm of the ground truth. This corresponds to a frame duration of 0.5-1 sec for typical clinical PET activity levels. Using four MLEM iterations with no subsets, a transaxial pixel size of 4 mm, a post-smoothing filter with 4-6 mm full width at half maximum, and averaging two or more frames to create the reference image provides an optimal set of parameters to produce accurate registrations while keeping the reconstruction and processing time low. CONCLUSIONS: It is shown that very short frames (≤1 sec) can be used to provide accurate and quick data-driven rigid motion estimates for use in an event-by-event motion corrected reconstruction.
Authors: Nicolas Costes; Alain Dagher; Kevin Larcher; Alan C Evans; D Louis Collins; Anthonin Reilhac Journal: Neuroimage Date: 2009-05-27 Impact factor: 6.556
Authors: Alexander M Grant; Timothy W Deller; Mohammad Mehdi Khalighi; Sri Harsha Maramraju; Gaspar Delso; Craig S Levin Journal: Med Phys Date: 2016-05 Impact factor: 4.071
Authors: Matthew G Spangler-Bickell; Mohammad Mehdi Khalighi; Charlotte Hoo; Phillip Scott DiGiacomo; Julian Maclaren; Murat Aksoy; Dan Rettmann; Roland Bammer; Greg Zaharchuk; Michael Zeineh; Floris Jansen Journal: IEEE Trans Radiat Plasma Med Sci Date: 2018-10-31
Authors: Murat Aksoy; Christoph Forman; Matus Straka; Stefan Skare; Samantha Holdsworth; Joachim Hornegger; Roland Bammer Journal: Magn Reson Med Date: 2011-03-22 Impact factor: 4.668
Authors: Ahmadreza Rezaei; Matthew Spangler-Bickell; Georg Schramm; Koen Van Laere; Johan Nuyts; Michel Defrise Journal: Phys Med Biol Date: 2021-09-13 Impact factor: 3.609
Authors: Kevin M Bradley; Timothy W Deller; Matthew G Spangler-Bickell; Floris P Jansen; Daniel R McGowan Journal: J Neurol Date: 2021-06-06 Impact factor: 4.849
Authors: Enette Mae Revilla; Jean-Dominique Gallezot; Mika Naganawa; Takuya Toyonaga; Kathryn Fontaine; Tim Mulnix; John A Onofrey; Richard E Carson; Yihuan Lu Journal: Neuroimage Date: 2022-03-04 Impact factor: 7.400
Authors: Eric Einspänner; Thies H Jochimsen; Osama Sabri; Bernhard Sattler; Johanna Harries; Andreas Melzer; Michael Unger; Richard Brown; Kris Thielemans Journal: EJNMMI Phys Date: 2022-03-03