Application of the CLEAN Algorithm to Three Dimensional Coded Aperture Imaging
Kevin Byard ()
Additional contact information
Kevin Byard: School of Economics, Faculty of Business, Economics and Law, Auckland University of Technology
No 2020-14, Working Papers from Auckland University of Technology, Department of Economics
Abstract:
An iterative technique for the removal of artifacts caused by the near field effects of a coded aperture imaging system is presented. The technique, we we call z-Clean, first locates high energy sources within a three dimensional field of view using a least squares method and then removes the artifacts using a method similar to that of the CLEAN algorithm used in radio astronomy, but instead operating in the detector shadowgram domain rather than the final image domain. Computer simulations where performed of observations of four point sources of different intensities and at different depths from the detector. Both a continuous detector of 1cm FWHM detection capability and a pixellated detector with 0.2cm square pixels were investigated using a Modified Uniformly Redundant Array coded aperture of element size 0.6cm. The efficacy of the z-Clean technique for artifact removal is demonstrated for both detector types for the three strongest sources of 100kBq, 50kBq and 10kBq using plane separations of 2cm, 1m, 0.5cm and 0.1cm, to leave only small ghosts lying up to aorund 2cm from the reconstructed source depth. For twenty trials of each observation, the three strongest sources are reconstructed no further than 0.7cm from the closest plane with many being from 0cm to 0.5cm for both detector types. The depth location for all three strongest sources using both detector types is no worse than 0.5cm from the actual source depth and in most cases is much better, being closer than 0.1cm for the strongest source at plane separations of 1cm, 0.5cm and 0.1cm. z-Clean was not able to remove the artifacts nor determine accurately the depth of the weakest source of 5kBq and in general sources that experience a phasing error are less accurately located although stil better than 0.5cm from the actual source depth for all such cases. The artifact removal and very good depth location come at the expense of an impact on the signal to noise ratio (SNR) of the sources. For the strongest source and using the continuous detector the SNR increases unexpectedly to give values higher than that for observations made only in the critical plane due to the ghosting of this source in other planes at different depths. For all other cases there is a decrease in SNR which is more makred for finer plane separations and for weaker sources.
Keywords: coded aperture; three dimensional imaging; tomography; gamma ray imaging; image processing (search for similar items in EconPapers)
Date: 2020-09
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.aut.ac.nz/__data/assets/pdf_file/0006/433491/working-paper-20_14.pdf (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:aut:wpaper:202014
Access Statistics for this paper
More papers in Working Papers from Auckland University of Technology, Department of Economics Contact information at EDIRC.
Bibliographic data for series maintained by Gail Pacheco ().