Jeff Fessler's image reconstruction opinions

as of ????

There are as many different methods for image reconstruction as there are researchers in the field. More in fact, because many researchers (myself included) have proposed multiple methods! A problem with archival papers is that by the time they appear, often the authors have already found better methods or variations. Also, because of the strong opinions (biases?) held by many researchers (myself included), there is often little opportunity for speculation and unsubstantiated opinion in refereed papers. This document is a collection of my current opinions on image reconstruction methods. These are subject to change, of course, but hopefully will be more current and more frank than what appears in my papers in the literature.

Some readers will disagree with these opinions. Since the web offers the ultimate opportunity for "equal time," I encourage dissenters to post their opinions in their own web sites, and I will gladly install pointers here to the dissenting opinions.

How to regularize?

Everyone knows that some form of regularization is required for image reconstruction, and the predominant choices seem to be stopping rules, sieves or post-smoothing, and penalty functions (aka priors). A serious problem with stopping rules is that different parts of the image converge at different rates, so you get object-dependent resolution and noise characteristics that are difficult to predict. (Barrett, Wilson, and Tsui (Phys. Med. Biol., 39:833-846, 1994) did an impressive job of analyzing ML-EM. Nevertheless, to my knowledge this analysis did not lead to a solution for the object-dependent resolution properties.) Sieves are often implemented as post-filtering, and often described as a formal regularization method by citing Snyder et. al. (IEEE-T-MI 6(3):228-238, 1987), but the equivalence between post-filtering and sieves hinges on a commutability condition (eqn. (12) of the above paper) which to my knowledge only has been shown to hold in the spatially-invariant Gaussian blur case used in that paper. But the real practical problem with post-filtering is that you first have to run your (unregularized) algorithm until "convergence," (which may be hundreds of iterations, unless you use one of the many acceleration methods that mostly have uncertain convergence properties) and then post-filter, otherwise some parts of the image may not have converged (e.g. Miller et. al., JNM 33(9):1678-1684, 1992). In contrast, there are now several "fast" algorithms for maximizing penalized-likelihood objective functions with proven global convergence. Also, the Washington University group that originally developed the method of sieves has found that MAP (or equivalently penalized likelihood) methods outperform the method of sieves (Butler et. al., IEEE-T-MI 12(1):84-89, 1993). So as you have guessed by now, I am a firm believer in the penalized-likelihood approach, and everything below applies only to algorithms for that case.

Estimators vs Algorithms

Many emission tomography papers discuss "image reconstruction algorithms" as if the algorithm is the estimator . This is partially true if you use stopping rules, since then the specific characteristics of the iterations determine the image properties perhaps more than the objective function does, since the user rarely iterates to convergence. With penalized objective functions, the estimator (as specified by the objective) determines the image properties, and the algorithm is merely a nuisance that is necessary for finding the maximizer that objective. As others have argued before me, the choice of the objective and the algorithm ideally should be kept separate, i.e., the objective should be chosen based on statistical principles, and then the algorithm should be chosen based on how fast it maximizes the chosen objective. All too frequently these two distinct issues have been blurred together. In the following I will attempt to keep the two choices separate.

Transmission Tomography

When transmission scans have low counts, or when they are significantly contaminated by "background" counts (such as random coincidences in PET or emission crosstalk in SPECT), both the FBP reconstruction method and penalized weighted least squares (PWLS) methods lead to attenuation maps with systematic biases due to the nonlinearity of the logarithm (Fessler, IEEE-T-IP, 4(10):1439-50, 1995). Therefore, I strongly recommend using penalized-likelihood objectives rather than PWLS objectives (or FBP) for noisy transmission measurements. (Penalized-likelihood methods avoid the logarithm by building the exponential into the statistical model.) Penalized-likelihood methods also allow you to properly account for additive background effects like cross-talk (SPECT) and randoms (PET) into the model.

What algorithm do I recommend for maximimizing penalized-likelihood objectives in transmission tomography? First, do not use the transmission ML-EM algorithm of Lange and Carson (1984, JCAT)!! That algorithm is very slow to converge, and requires inordinate computation per iteration (and is also not guaranteed to be monotonic if the approximate M-step is used). I have published several papers describing faster converging algorithms for this problem. All of these papers have been made obsolete by our most recent method, based on Parabaloidal Surrogates. This method was presented by Hakan Erdogan at the 1998 ICIP conference, and is currently in review for IEEE T-MI. Email me if you would like a preprint. This latest method is guaranteed to be monotonic, even for the nonconvex case, and the CPU time per iteration is comparable to that of PWLS methods. So there is no longer any computational advantage of using PWLS over penalized likelihood for transmission tomography. (The algorithm as also available in ASPIRE.)

Emission Tomography

If you have truly Poisson data (which is typically the case in SPECT, and sometimes the case in PET), then again I recommend a penalized-likelihood objective rather than the penalized data-weighted least squares (PWLS) objective, since data-weighted least squares again leads to biases at low count rates (Fessler, IEEE-T-IP, 5(3):493-506, 1995). If your PET data has been precorrected for random coincidences, then I recommend the shifted Poisson method (Yavuz and Fessler, Med. Im. Anal., to appear). If you have emission data that has been unavoidably "corrupted" by other correction methods, interpolations, etc. (such as Fourier rebinning - see papers by Comtat and Kinahan), and is no longer Poisson, then you may as well use PWLS, since the algorithms for PWLS typically converge a little faster than the algorithms for penalized likelihood.

Emission Penalized Likelihood

Currently, I recommend the SAGE algorithm (Fessler and Hero, IEEE-T-IP 4(10):1417-29, 1995) for maximizing emission penalized-likelihood objectives, specifically the PML-SAGE-3 version of the algorithm, which supersedes the earlier SAGE algorithms (Fessler and Hero, IEEE-T-SP 42(10):2664-77, 1994). However, I will be presenting a Parabaloidal Surrogates algorithm at NSS/MIC 1998 that is comparable or slightly faster than SAGE, but is easier to implement, so you may prefer that method. Email me for preprints.

Penalized Weighted Least Squares

If you decide you would rather use the PWLS objective (for reasons of simplicity, non-Poisson data, or possibly algorithm speed), then I again recommend a coordinate ascent algorithm. A nice thing about coordinate ascent is that not only does it converge very fast (if you give it a reasonable starting image like an FBP image, not a uniform image, see the nice frequency-domain analyses of Sauer and Bouman (IEEE T-SP, 41(2):534-548, 1993)), it also handles nonnegativity trivially. But if nonnegativity is not an issue, then the conjugate gradient algorithm is very attractive, and if you precondition it well, then it appears to converge at roughly the same rate as (properly under-relaxed) coordinate ascent, and has the advantage of being more parallelizable. See the nice work of Delaney and Bresler (ICASSP 95) and the nice recent paper by Mumcuoglu and Leahy in IEEE T-MI, who also address nonnegativity.

Penalty Functions

Giving any opinions about this is an optimal way to alienate the greatest number of people in the shortest amount of time. So here goes. First decision: convex or non-convex? Bouman and Sauer (IEEE-T-IP 3(2):162-177, 1994) showed that non-convex penalties lead to estimates that are discontinuous functions of the measurements. This means that it is possible that changing the measurements by even just one photon count could change the image significantly if you use a nonconvex penalty. To me this seems unacceptable for medical imaging, although one has to admit that the reconstructed images (of piecewise constant phantoms, usually) look very nice. Time may tell otherwise, but I am mostly sticking to convex penalties for now. Second decision: quadratic or non-quadratic? Quadratic penalties (with 1st or 2nd-order neighborhoods) yield Butterworth-filter-like resolution properties, so the images have reduced noise relative to FBP, but comparable resolution characteristics. I think this makes quadratic penalties well suited to the first couple rounds of clinical studies comparing FBP with statistical methods. The non-quadratic penalties have noise texture that to me looks very different, and may take more time for the clinicians to adapt to. But for quantitative studies (non-visual), if the object under study is really piecewise constant or maybe piecewise smooth, then there is no doubt that non-quadratic penalties give a better bias/variance tradeoff. The questions unanswered are whether brains and tumors and such are piecewise constant, and if not, how the (different) biases of quadratic and non-quadratic penalties would compare.

For all penalties, there is an interaction between the Poisson measurement covariance and the penalty that leads to nonuniform spatial resolution. I have proposed a modified penalty makes the spatial resolution nearly uniform for nearly spatially-invariant tomographs (i.e. 2D PET), and I now use it routinely. But I have not investigated spatially-variant systems like SPECT enough to comment conclusively. My intuition is that the proposed modified penalty will somewhat help to improve the resolution uniformity, but there are probably better modifications. I have more work to do there. Web Stayman and I have recently developed a new penalty function that leads to more uniform spatial resolution, presented at ICIP 98. Email if you would like preprints of this paper.

If you would rather have uniform noise than uniform resolution, then in principle you could use Sterling-type (co)variance approximations to predict the variance as a function of the smoothing parameter, and then adjust the parameter spatially to achieve your desired noise level. But that would actually be pretty complicated, so finding a simpler way is an open problem.

Differing opinions?

Email "fessler AT umich DOT edu"
Back to Fessler home page