Residual 3D convolutional neural network to enhance sinograms from small-animal positron emission tomography images

Positron emission tomography (PET) has been widely used in nuclear medicine to diagnose cancer. PET images suffer from degradation because of the scanner’s physical limitations, the radiotracer’s reduced dose, and the acquisition time. In this work, we propose a residual three-dimensional (3D) and c...

Full description

Saved in:
Bibliographic Details
Other Authors: Rodríguez, Leandro José, Ochoa Domínguez, Humberto, Vergara Villegas, Osslan Osiris, Cruz Sanchez, Vianey Guadalupe, Polanco Gonzalez, Javier, Sossa, Juan Humberto
Format: Artículo
Language:en_US
Published: 2023
Subjects:
Online Access:https://doi.org/10.1016/j.patrec.2023.05.005
https://www.sciencedirect.com/science/article/abs/pii/S0167865523001320
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Positron emission tomography (PET) has been widely used in nuclear medicine to diagnose cancer. PET images suffer from degradation because of the scanner’s physical limitations, the radiotracer’s reduced dose, and the acquisition time. In this work, we propose a residual three-dimensional (3D) and con- volutional neural network (CNN) to enhance sinograms acquired from a small-animal PET scanner. The network comprises three convolutional layers created with 3D filters of sizes 9, 5, and 5, respectively. For training, we extracted 15250 3D patches from low- and high-count sinograms to build the low- and high-resolution pairs. After training and prediction, the image was reconstructed from the enhanced sino- gram using the ordered subset expectation maximization (OSEM) algorithm. The results revealed that the proposed network improves the spillover ratio by up to 4.5% and the uniformity by 55% compared to the U-Net. The NEMA phantom data were obtained in a simulation environment. The network was tested on acquired real data from a mouse. The reconstructed images and the profiles of maximum intensity projection show that the proposed method visually yields sharper images.