Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Towards guidelines to harmonize textural features in PET: Haralick textural features vary with image noise, but exposure-invariant domains enable comparable PET radiomics

  • George Amadeus Prenosil ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    george.prenosil@insel.ch

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Thilo Weitzel,

    Roles Conceptualization, Formal analysis, Methodology, Project administration, Software, Supervision, Validation, Writing – original draft, Writing – review & editing

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Markus Fürstner,

    Roles Conceptualization, Validation, Writing – review & editing

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Michael Hentschel,

    Roles Conceptualization, Validation, Writing – review & editing

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Thomas Krause,

    Roles Conceptualization, Funding acquisition, Project administration, Resources, Writing – review & editing

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Paul Cumming,

    Roles Writing – review & editing

    Affiliations Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland, School of Psychology and Counselling and IHBI, Queensland University of Technology, Brisbane, Australia

  • Axel Rominger,

    Roles Funding acquisition, Project administration, Resources, Supervision, Writing – review & editing

    Affiliation Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Bernd Klaeser

    Roles Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing

    Affiliations Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland, Department of Radiology and Nuclear Medicine, Cantonal Hospital Winterthur, Winterthur, Switzerland

Abstract

Purpose

Image texture is increasingly used to discriminate tissues and lesions in PET/CT. For quantification or in computer-aided diagnosis, textural feature analysis must produce robust and comparable values. Because statistical feature values depend on image count statistics, we investigated in depth the stability of Haralick features values as functions of acquisition duration, and for common image resolutions and reconstructions.

Methods

A homogeneous cylindrical phantom containing 9.6 kBq/ml Ge-68 was repeatedly imaged on a Siemens Biograph mCT, with acquisition durations ranging from three seconds to three hours. Images with 1.5, 2, and 4 mm isometrically spaced voxels were reconstructed with filtered back-projection (FBP), ordered subset expectation maximization (OSEM), and the Siemens TrueX algorithm. We analysed Haralick features derived from differently quantized (3 to 8-bit) grey level co-occurrence matrices (GLCMs) as functions of exposure E, which we defined as the product of activity concentration in a volume of interest (VOI) and acquisition duration. The VOI was a 50 mm wide cube at the centre of the phantom. Feature stability was defined for df/dE → 0.

Results

The most stable feature values occurred in low resolution FBPs, whereas some feature values from 1.5 mm TrueX reconstructions ranged over two orders of magnitude. Within the same reconstructions, most feature value-exposure curves reached stable plateaus at similar exposures, regardless of GLCM quantization. With 8-bit GLCM, median time to stability was 16 s and 22 s for FBPs, 18 s and 125 s for OSEM, and 23 s, 45 s, and 76 s for PSF reconstructions, with longer durations for higher resolutions. Stable exposures coincided in OSEM and TrueX reconstructions with image noise distributions converging to a Gaussian. In FBP, the occurrence of stable values coincided the disappearance of negatives image values in the VOI.

Conclusions

Haralick feature values depend strongly on exposure, but invariance exists within defined domains of exposure. Here, we present an easily replicable procedure to identify said stable exposure domains, where image noise does not substantially add to textural feature values. Only by imaging at predetermined feature-invariant exposure levels and by adjusting exposure to expected activity concentrations, can textural features have a quantitative use in PET/CT. The necessary exposure levels are attainable by modern PET/CT systems in clinical routine.

Introduction

Positron emission tomography in conjunction with computed tomography (PET-CT) is an imaging modality that quantifies radiotracer uptake in the living organism. With CT-based attenuation correction, and correction for scatter, randoms, and dead time, PET depicts absolute tracer uptake in units of concentration. In addition, the PET/CT system`s spatially resolved measurements reveal textural features, yielding a metric for quantifying an imaged object’s structure [1]. Textural feature analysis increases the information retrievable from PET/CT images, which is useful for characterizing and classifying tissue and lesions according to differences in morphology [2]. Adding texture analysis to a PET/CT readout can thus improve diagnostic sensitivity and specificity for certain tumor entities [3]. Notably, textural features are exploited as a surrogate measure of tissue metabolic heterogeneity [4], which can be an index of tumor malignancy, for example in the case of thymic epithelial cancer [5]. Alternately, changes in textural feature values depict response to treatment in esophageal cancer [6]. Expanding the feature space of PET/CT images makes texture analysis useful for pattern recognition techniques [7] and computer-assisted diagnosis (CAD).

Among the various mathematical methods to quantify image texture, partial gray level co-occurrence texture features are the best-established, having been introduced in 1973 by Haralick et al. [8] for the automated segmentation of satellite images. Sometimes also referred to as second order statistics, Haralick features have since found their way into radiomics [9] and subsequently PET/CT imaging [7]. However, images—and hence the measured tracer uptake values—typically differ between the various clinical PET/CT systems, even when imaging the very same object [10]. Distinct instrumentation and differences in image reconstruction algorithms impart characteristic imaging properties to PET/CT systems [11]. Resultant differences in imaging properties limit the comparability of datasets acquired on different PET/CT systems or with different acquisition protocols, and hamper the use of quantitative tracer uptake values in clinical routine or in multicenter clinical trials [12].

[13,14] Image noise, with its texture-like manifestations, varies especially with frame duration, tracer uptake, and the sensitivity of a given PET/CT system, which hampers comparison of datasets.

Despite certain efforts to quantify image noise and its effects on tracer uptake measurements [1517] and image quality [18], the optimal image exposure remains unresolved for many clinical situations, and especially so in the context of textural feature values. These values must not only be stable in longitudinal, repeated measurements in a patient [19], but must be comparable between different PET/CT systems, if they are to be of diagnostic or prognostic value. The comparability problem of textural feature values is indeed analogous to the more widely recognized problem of variability in measured tracer uptake values.

Several works have examined textural feature stability [20], either under differing image acquisition and reconstruction conditions [21,22], in terms of grey level quantization [23], or under multicenter conditions [24], to identify the least variant features. However, the contribution of image count statistics, a known driver of variability in imaging, remains largely overlooked. Although, some studies show effects of noise on image texture [2527], others have reported insignificant effects of image statistics on textural features [28,29]. In fact, there has hitherto be no systematic study of the progression of textural feature values with changing photon count. In most literature reports, the variability of textural features arising from image noise is usually described in terms of variance [30] and within the same exposure regime [31], without reporting on possible domains with stable feature values. This implies that radiomic values are not readily comparable between different PET/CT systems with different sensitivities and imaging properties.

The voxel (second order) neighborhood statistics of Haralick features must depend on their underlying first order statistics, i.e. the intensity distribution of the voxel image. At the detector level, Shot noise [32] will dominate in PET/CT and the signal-to-noise ratio (SNR) will be proportional to the square root of the imaged counts. Shot noise follows a Poisson distribution, but the noise distribution in the final PET/CT image will depend on the chosen image reconstruction method and its parameters. For examples, PET/CT images acquired with sufficient counts and reconstructed with filtered back projection (FBP) approximate Gaussian statistics [33], whereas ordered subset expectation maximization (OSEM) [34,35] or point spread function (PSF) based resolution modeling tend to resemble log-normal [36,37] or gamma distribution statistics [33]. Furthermore, with OSEM or PSF-based algorithms, the covariance matrix of a given image region will depend on the very region itself, e.g. the mean image [36]. This non-linearity renders mathematical predictive approaches to noise behavior for such reconstructions difficult and ill-posed. Image noise propagates linearly through the FBP reconstruction [38] or non-linearly, in an iteration dependent manner through OSEM- [36,39] and PSF-based reconstructions [40]. Therefore, noise increases the apparent roughness of the final image [17,41] differentially for different image reconstructions.

It follows that PET/CT acquisition and reconstruction protocols delivering consistent and comparable features values require prior knowledge of the dependency of textural features on the count statistics and on imaging properties. Only then can textural features become comparable across sites using diverse PET/CT systems, which is a necessary condition for quantitative usage of texture metrics in multi-center studies. Given this background, our aim was to investigate experimentally the count statistics dependency of first order statistical and Haralick features commonly used in PET/CT. For practical applications, we elaborate a methodology for identifying exposure domains with minimal variability in textural features. We apply our methodology on Haralick textural features measured in seven different image reconstructions, but our approach is generalizable to other image reconstructions and feature definitions.

To test feature stability, we obtained PET/CT measurements of a homogenous, solid-state Ge-68 phantom–just as used for daily quality assurance (QA) tests—in our clinical PET/CT system, with repeated measurements for durations ranging from three seconds to three hours. The product of acquisition times and phantom activity covered a wide exposure range. -We measured the selected textural features within the same central region of the cylindrical phantom for seven clinical image reconstruction protocols, thus testing our hypothesis that textural feature values are highly variable depending on image noise, resolution, and reconstruction procedures. Additionally, we looked for criteria that convey stability to Haralick feature values in different image reconstructions. In so doing, we present a simple and automated way to identify exposures with stable image features for any given PET/CT system.

Material and methods

Data acquisition

Phantom measurements of varying exposure were acquired and reconstructed on a Biograph mCT-X 128 (Siemens Medical Solutions USA, Knoxville, TN) using the same Ge-68 solid-state phantom as used for daily quality assurance (QA) (Eckert & Ziegler Isotope Products, Valencia CA, USA). This cylindrical phantom had an inner diameter of 200 mm, cylinder wall thickness of 6.5 mm, and an active volume of 8407 ml of 68-Ge homogeneously distributed in a hardened epoxy matrix.

Our use of the QA phantom imparts an easily reproducible methodology in clinical settings. We axially aligned and centered the phantom in the PET/CT system’s field of view (FOV) as specified for the daily QA, and imaged only a single bed position centered on the phantom. We acquired 20 PET/CT measurements with 25 different frame durations increasing from three seconds to a final 10861 seconds at a factor of per frame. All data were reconstructed with CT-based attenuation correction, with scatter correction, and in the time-of-flight mode (TOF).

In the contexts of image noise and comparability, we prefer exposure to acquisition duration as a measure for the actual number of available decays. We define exposure as the product of activity concentration in a volume of interest (VOI) and acquisition duration or, in the case of relevant decay times, the integral of activity concentration over the acquisition time. In addition, by introducing a factor representing the relative sensitivity of another PET/CT system or e.g. differing decay branching ratios of other positron emitters, results can easily be rescaled to match differing systems and differing acquisition conditions.

Image reconstruction

Seven different image reconstruction protocols combined three different image resolutions with three different reconstruction methods. All images were reconstructed with isotropic voxels and with post-reconstruction Gaussian filtering. The full width at half maximum (FWHM) of the Gaussian post-reconstruction filter ranged from 1 to 4 mm. Standard-resolution (SR) images had voxels of around 4 mm isometric spacing, whereas high-resolution (HR) images had 2 mm spaced voxels. Because (and contrary to FBP and OSEM) the Siemens TrueX algorithm enables higher resolution PET/CT, so called ultra-high resolution (UHR) voxels of around 1.6 mm isometric spacing were paired with said algorithm [42] in a single protocol. The three different reconstruction algorithms used were FBP, OSEM, and TrueX, the latter being a PSF-based resolution-modeling algorithm specific to Siemens PET/CT systems. Table 1 provides an overview of the seven different protocols and their exact parameters.

Data analysis

A cuboid of around 50 mm edge length, which was centered in the daily QA phantom, defined the VOI from which all first and second order statistics were calculated (Fig 1A and 1B). The actual cuboid edge length in a given spatial direction was adjusted to the next even voxel count, yielding arrays (c) with nx*ny*nz = N voxels. (cf. Table 1). Exposure EAcq in the cuboid VOI was defined as the product of the known activity concentration AC in that region and acquisition duration, tAcq 1

Our present use of exposure instead of acquisition duration for most graphs and measures makes results in this work independent of the actual AC, and readily transferable to other ACs. With 270.8 d half-life, the Ge-68 source retained a nearly constant activity concentration during the study, but decay would have to be considered in studies with short-lived nuclides.

The PET/CT data was automatically analyzed using a multi-paradigm software [43] written in-house. This software detected the phantom and calculated all statistics in a cuboid VOI placed automatically in the center of the phantom.

Grey level co-occurrence matrices.

Six different grey level co-occurrence matrices (GLCMs) with 8 (3 bit), 16 (4 bit), 32 (5 bit), 256 (8 bit), 128 (7 bit) and 64 (6 bit) rows and columns were calculated from the central cuboid VOI, where each entry in the matrix represented the relative occurrence frequency of orthogonally adjacent voxel pairs with intensities ® and j. The number of rows and columns corresponded to G, the number of gray levels used. The relative occurrence frequency was calculated by dividing the number of ®,j-voxel pairs (p) by the total number of voxel pairs ® found in the cuboid VOI.

2

In this work, we used only orthogonal nearest neighbor voxels, with a displacement d of one voxel and displacement angled θ of 0° and 90°. Diagonal voxels with d > 1 were excluded from consideration. We are aware of the GLCM’s scale and direction sensitivity arising from use of different d and θ values, but assuming isotropic image noise, no directional preference is to be expected. Furthermore, by using solely a nearest neighbor geometry and not a fixed distance d, the spatial sensitivity of the GLCMs scaled with the spatial sampling density of the PET/CT system and the chosen image resolution. The large number of voxels analyzed in the cuboid VOIs (cf. Table 1) justified the use of up to 256 gray levels, without producing information-free voids in GLCMs. This in turn enabled the statistical analysis of textural features using Haralick’s methods.

Grey level mapping in GLCMs was performed as described in Tixier et al. [6]. This method resampled the gray levels found in a VOI to the finite range of the GLCMs, linearly mapping the lowest intensity in the VOI to the first gray level and the highest value to the last gray level of a GLCM.

First order statistics.

In addition to the cuboid’s mean μc and standard deviation σc, nine voxel value statistics were calculated from the central cuboid array c, as summarized in Table 2. These additional nine metrics, here called first-order statistics, often find use along with texture as descriptors of PET/CT images. In this work, we used Eqs 4–8 and 11 as descriptors for normality of the AC distribution in the VOI.

thumbnail
Table 2. First order statistical features used for texture analysis.

https://doi.org/10.1371/journal.pone.0229560.t002

Haralick features (second order statistics).

Eleven different Haralick features were calculated from the GLCMs as first described in Haralick et al. [8] and later summarized by Albregtsen [45]. Given below are the mathematical definitions of these texture features (Table 3).

The mathematical terms and definitions used inside the feature equations are given below, starting with the mean in x- and y- direction of the GLCM matrix: 23 24

Standard deviation in x- and y-direction of the GLCM matrix: 25 26

Greyscale sum and difference vectors Px+y and Px-y 27 28

Mode was calculated in x direction (Modex) as an additional statistical descriptor of GLCMs.

Feature values F obtained from within the aforementioned VOI were plotted logarithmically against exposure E and interpolated into continuous curves by locally estimated scatterplot smoothing (LOESS) curves using IGOR pro (Wavemetrics, Lake Oswego, Oregon 97035, USA), which was also used to generate graphs and tables. The LOESS was conducted with a sliding smoothing window with a width of thirteen measurement points. To equalize the contribution of geometrically increasing exposure values, we ran the LOESS on log-log data. The LOESS curves were then used to identify stable maxima and minima or plateaus, where the onset of a plateau is defined when the relative slope in LOESS curve became less than 5%/E, with E in units of kBq/ml*s.

29

Results

Noise distribution

To understand the impact of exposure upon image texture, PET/CT measurements of the daily GE-68 QA phantom were recorded at increasing acquisition durations. The phantom AC was 9.54 kBq/ml at the time of measurement, and thus the acquired exposures ranged from 28.6 kBq/ml*s to 103.6 MBq/ml*s with the chosen durations. Fig 1A illustrates how the apparent coarseness of the HR FBP images decreases with increasing acquisition duration. Fig 1B shows the same effect, albeit on a different resolution scale, for UHR PSF images. As examples, Fig 1C and 1D show the activity distribution histograms of the cuboid VOI in the cases of HR FBP and UHR PSF reconstructions. While activity histograms reconstructed with FBP tended be wider than were those from PSF reconstructions, the former were more Gaussian in shape, even at brief acquisitions. In addition, FBP data displayed negative voxel values and required longer acquisition before their activity distributions moved away from zero than did OSEM (not shown) or PSF data. As expected, the spatial sampling density showed similar effects, with SR data having narrower and more Gaussian histograms than did HR or UHR data (not shown). The measured average activity concentration in the VOI was unaffected by exposure in all examined reconstructions (S1 Fig).

thumbnail
Fig 1. Noise in PET/CT images at different exposures.

Shown are examples of four differently exposed HR FBP (a) and UHR PSF (b) images, using the daily QA phantom. Images are maximum intensity projections (MIPs) in the axial direction of the PET system with black set to naught and white set to the maximal intensity found in the PET image. The green rectangle marks the VOI, where all the calculations were performed. The associated noise histograms of the HR FBP (c) and UHR PSF (d) measurements are depicted below, showing the measured AC distribution in the VOI after four different acquisition durations. Nvoxel denotes the number of analyzed voxels in the VOI. Black insets are magnified histograms of the longest acquisition. Color legend of acquisition durations applies to both (c) and (d).

https://doi.org/10.1371/journal.pone.0229560.g001

Exposure dependency of first order statistical features

To reveal how noise propagates through an image reconstruction process, we calculated nine different statistical voxel value metrics from the central cuboids (Fig 2 and Eqs 3–11). When plotted against exposure, these measurements reflected the above findings in the histograms: While CV, crest factor, and normalized entropy followed the expected power law for all seven reconstruction protocols, excess kurtosis and skewness were constant for FBP but varied in OSEM and PSF data. In the latter datasets, skewness resembled CV with respect to exposure, but excess kurtosis initially followed a power law, only to stabilize around zero after exceeding some minimum exposure. SR data had no excess kurtosis, even at very low exposure values. The presence of negative values in FBP images as seen in Fig 1C caused a plateau to appear in the crest factor at low exposures. Negative values ceased to appear after 30 s in SR FBP and 679 s acquisition time in HR FBP data, translating to exposures of 286 kBq/ml*s and 6476 kBq/ml*s, respectively.

thumbnail
Fig 2. First order statistics of image noise measurements as a function of exposure and grouped according to different reconstruction protocols.

https://doi.org/10.1371/journal.pone.0229560.g002

While kurtosis and skewness are accepted indicators for normality, we additionally used Lillifor’s p and the goodness of fit of a Gaussian to the noise distribution. Lillifor’s p remained close to unity for FBP data irrespective of exposure but increased from an initially low magnitude (10E-10) for OSEM and PSF data. Here, p showed a sigmoidal approach towards a plateau at unity, in the same domain of exposure where excess kurtosis disappeared. Its associated test statistic D is the supremum between the sample cumulative distribution and a normal distribution of equal mean and variance. Log plots revealed D decreasing with exposure and following a power law, but with parallel ordinate shifts for the seven different reconstruction protocols. There was less pronounced exposure dependency of D in FBP data, but the correlation was still visible.

Similarly, the Gauss fit NRMSD decreased with a power law with increased exposure for all seven reconstructions. As with the other metrics used, the AC distribution in FBP data more closely resembled a Gaussian distribution at lower exposures than required for OSEM and PSF data to attain a Gaussian. The mode/mean ratio reflected the increased skewness and non-Gaussian shape of the AC distribution at low exposures, while approaching unity in congruence with the plateauing of the Gauss fit NRMSD.

Exposure dependency of grey level co-occurrence matrices

As a next step, we calculated GLCMs with different numbers of greyscales, mapping the PET data over a range extending between the respective minimum and maximum activity concentrations found in the cuboid VOI. Fig 3 shows examples of the GLCM dependency on acquisition duration for different HR reconstructions before greyscale mapping. Similar to the noise histograms noted above, these FBP GLCMs were broader, more Gaussian and less skewed than were OSEM or PSF GLCMs. The latter two types of GLCM showed an alignment on the diagonal at high exposures, with more skewness and some dispersion at low exposures. When shown in units of activity concentration, all GLCMs had exposure-invariant average values (μx and μy; Eqs 23 and 24). By contrast, Modex approached μx only at higher exposures, while σx (Eq 25) showed the expected power law behavior.

thumbnail
Fig 3. Overlaid examples of 8 bit GLCMs in units of kBq/ml.

GLCMs calculated from HR datasets and reconstructed with either with FBP, OSEM, or TrueX, PSF. Five selected acquisition durations are shown at their location in respect to absolute activity concentrations. For visualization (top). GLCM maxima are normalized to unity and are displayed with a logarithmic color scale. Acquisition times are noted besides the PSF GLCM in the respective color.

https://doi.org/10.1371/journal.pone.0229560.g003

After greyscale mapping, the shape of GLCMs remained, but only GLCMs from FBP had a constant μx, Modex, and σx. In OSEM and PSF reconstructions, these values showed a sigmoid function (Fig 4). Comparing Figs 2 and 3 suggests that greyscale mapping produced GLCMs of similar extension with regard to the grey scale count, because only using values occurring inside the VOI. Especially in the case of FBP, this led to GLCMs with a Gaussian cross-section of similar extent, while PSF based reconstructions still had GLCMs of varying shape and extent. Mapping the GLCMs in greyscale units maintained the general shape of the matrices, but revealed how μx, Modex, and σx moved towards higher values with increasing exposure in OSEM and PSF reconstructions, while remaining constant in FBP reconstructions (Fig 4).

thumbnail
Fig 4. Overlaid examples of 8 bit GLCMs in units of greyscales.

GLCMs calculated from HR datasets and reconstructed with either with FBP, OSEM, or TrueX, PSF. Five selected acquisition durations are shown at their location in respect to absolute their greyscale values. GLCM maxima are normalized to unity and are displayed with a logarithmic color scale. Acquisition times are noted besides the PSF GLCM in the respective color.

https://doi.org/10.1371/journal.pone.0229560.g004

Exposure dependency of Haralick features

From the full range GLCMs, 11 Haralick or so called second-order statistical features were calculated and plotted against exposure (Eqs 12–22). The resulting scatterplots were smoothed by the LOESS method to obtain differentiable curves for every feature (Fig 5). Most feature-reconstruction combinations were variant at lower exposures before reaching an invariant plateau. Feature values from FBP stayed mostly constant over all exposures. The feature value curves behaved similarly in shape regardless of GLCM bit size. Nevertheless, varying the GLCM bin size resulted in differing absolute feature values; something that had been reported before [46].

thumbnail
Fig 5. Exposure dependency of Haralick texture features according to different image resolution and reconstruction algorithm calculated from 8-bit GLCMs with 256 grey levels.

Measurement points are shown in the first graph but omitted in the following ones for clarity, and only the interpolated Loess curves are shown. Loess confidence interval are not shown. In the ASM panel, a clinical relevant exposure range is shown in grey, taken from Krarup et al. [47]. Legend applies to all.

https://doi.org/10.1371/journal.pone.0229560.g005

Fig 6 shows the magnitude of variability between the lowest and highest values found in each textural feature calculated from the different GLCMs, as well as those found in first order statistics. While the highest variability of features extended over three orders of magnitude, the variability was almost similar between GLCMs regardless of their quantization levels (Figs 6 and S2A). Lowering the number of GLCMs greyscales led overall to a significant decrease in feature value range, but the effect size was negligible, with 0.013 to 0.028 Bel (S2B Fig).

thumbnail
Fig 6. Magnitudes of textural feature variability.

Shown is the absolute range between the lowest and highest values in units of Bel found in the texture features across all examined exposures calculated 8-bit GLCMs (a), from 7-bit GLCMs (b), from 6-bit GLCMs and were also found in first order statistics (d). Lilliefor’s p varied up to 12 orders of magnitude and is therefore not shown to scale. Color Legend in (a) applies to all.

https://doi.org/10.1371/journal.pone.0229560.g006

We also observe that SR FBP generally provided for more stable features than did HR OSEM and HR PSF reconstructions. Among the individual features, ENT, CORR, SENT, and DENT were the most stable, whereas SHADE and PROM were the most unstable.

First order statistics, which are sometimes also used as image descriptors, varied up to three orders of magnitude with exposure (S1C Fig), with exception of Lilliefor’s p, here used only as a Gaussian metric. Contrary to Haralick features, first order statistics mostly follow a power law concerning exposure and do not attain stable domains per se (Fig 2). Therefore, highest possible exposures should be sought when using first order statistics in PET/CT.

In SR reconstructions stable plateaus in Haralick features occurred at lower exposures than in HR reconstructions. Some feature-reconstruction combinations were stable over the entire exposure range (Marked with asterisks in Fig 7). Still, other feature-reconstruction combinations, such as SHADE in SR OSEM, attained stability distinctively later than the median exposure in their group. In terms of acquisition duration, median time to stability did not differ significantly between the differently sampled GLCMs (Fig 7). For example, when using 8-bit GLCM, median time to stability was 16 s for SR and 22 s for HR FBP, 18 s for SR and 125 s for HR OSEM, and 23 s, 45 s, and 76 s for SR, HR and UHR PSF reconstructions, respectively. In addition, we note that the occurrence of stable feature values in FBP reconstructions coincided the disappearance of negatives image values as occurs above certain exposures. In OSEM and PSF reconstructions stability coincided with the approach of noise histograms towards normality.

thumbnail
Fig 7. Onsets of stable plateaus in features calculated from differently quantized GLCMs with median exposure as a red line.

*Missing values indicate stability over the entire exposure range. Textural feature order applies to all sets of image reconstructions.

https://doi.org/10.1371/journal.pone.0229560.g007

Because texture feature values also depend on the GLCM quantizing scheme [46], we studied the effects of differently quantizing GLCMs on feature stability. In these calculations, we used the full range of activity concentrations delivered by the PET/CT system for quantizing GLCMs (S3 Fig) rather than normalizing the grey level range to the intensities in the VOI. Without this greyscale mapping, long-term exposure stability failed to occur, although there emerged what we term “islands of stability” (S4 Fig). The resulting feature value-exposure curves were highly non-monotonic, and often had no distinct plateaus (S5 Fig). Variability was up to two orders of magnitude higher (S6A Fig). However, islands of stability occurred at similar exposures as the onsets of plateaus described earlier (S6B Fig). See Supporting Material and Methods (S1 Text) and Supporting Results (S2 Text) for details.

In cases where the greyscale mapping was restricted to a fixed bin size scheme with a lower bound on zero (c.f. S1 Text) as suggested by Presotto et al. [26], variability inflated to four orders of magnitude, with almost no asymptotic plateaus evident in the feature value-exposure curves (S7 Fig).

Discussion

This work shows how exposure and hence image noise affects Haralick texture features on PET/CT measurements obtained with the homogenous daily QA phantom. By varying exposure (Eq 1) over a wide range and including values relevant for clinical practice, we report that feature values varied by more than two orders of magnitude in PET/CT images of an otherwise absolutely homogeneous phantom. Additional variability arose due to the different image reconstruction methods, with FBP producing the most robust Haralick features and non-linear iterative image reconstructions (PSF and OSEM) producing the largest variability. Consequently, image resolution correlated inversely with the exposure levels needed for obtaining stable Haralick feature values. Most feature-reconstruction combinations showed monotonic curves with stable plateaus (Fig 5) when plotted against exposure. Greyscale mapping produced GLCMs with congruent positions within the matrix, making full use of the available matrix size (Fig 4). On the other hand, full-range GLCMs drifted across the matrix according to the grey scale range occurring in the image (S2 Fig). This additional degree of freedom could explain the non-monotonic curves and the larger variability of full range GLCMs (S4 and S5 Figs and S2 Text).

The large variability in feature values is a function of how PET-CT image noise propagates into the final image: While FBP tended to produce Gaussian intensity distributions in homogenous VOIs in the phantom regardless of exposure, OSEM and TrueX reconstructions approached normality only with increasing photon count. The shape of these first order statistical distributions subsequently defined shapes of the (second order) GLCMs from which all Haralick texture features were eventually calculated. Consequently, once the first order distributions attained normality, most of the feature values became exposure-invariant. In this respect, the median exposure for stable plateaus (Fig 7) coincided with the exposures needed to bring Lilliefor’s p to unity (Fig 2).

Textural features in medical images occur as a function of the spatial distribution of the signal intensity in the structure. At the same time, this signal distribution is contaminated with noise. When signal is low within a given structure, the textural features may not emerge within a reasonable acquisition time, given the constraints of clinical practice. Moreover, regulations of radiation protection (diagnostic reference levels) limit the dose of injected tracer. These factors contribute to the difficulties in comparing texture features between sites.

The exposure regime where Haralick features become stable is within a clinically relevant range. Looking at the minimal mean uptake of 4 kBq/ml (calculated from SUVmean values and de-corrected for decay) encountered, for example, in lung cancer [47] and a median exposure of 729 kBq/ml for HR PSF reconstructions (6-bit GLCM, Fig 7), the minimal exposure to obtain more than 50% of exposure-invariant Haralick features is 182 s. This exceeds the 120 s used by the authors of the study [47], and furthermore, the features ASM, CON, VAR, and AVER remain influenced by exposure. From the minimal and maximal SUVmean in the study [47], we calculated and displayed an approximate clinical exposure range in the ASM panel of Fig 5, the strongest study correlate. One might ask, if better exposure control would have decreased the number of patients needed for the study or given better results to other Haralick features.

In dynamic clinical studies, acquisitions times for time activity curves are often very brief, and feature analysis is not regularly done. Nevertheless, let us assume a tracer uptake of around 8 kB/ml as seen in a single lesion (de-corrected for decay) in the work of Schwartz et al. [48]. Our results in Fig 7 then suggest minimal acquisition durations of 27 s and 89 s respectively would be required to obtain stable values on our PET/CT system. Such brief acquisition times bring feature analysis of dynamic acquisitions into the realm of the possible in a clinical setting, even though the minimal acquisition times still exceed the 10 s intervals used by Schwartz et al at the beginning of their time activity curves. As a matter of fact, with the somewhat lower sensitivity of their PET/CT system [49], the exposures from their 10 s intervals are approaching those from our briefest acquisition duration. Furthermore, in the case of UHR PSF reconstructions with 25 MBq/ml*s exposure required to obtain stable VAR values (Eq 17), the acquisition time soars to 52 minutes, assuming no increase in injected dose is sought.

Greyscale mapping, by giving stable plateaus (Fig 5) and lowering variability [24] (Fig 6), improves the stability and comparability of all feature values, making it a prerequisite for any feature analysis. Furthermore, stability occurs at lower exposures with greyscale mapping (Figs 7 and S5). However, this probably only holds true in a perfectly homogenous VOI, such as used in this work: Figs 4 and S5 and S7, and the work of Presotto et al. [26] hint that the quantization distance between grey level groups can introduce new variability. Such a situation might occur were lesions of similar texture are found within differing background ACs. It is therefore unsurprising that studies without greyscale mapped GLCMs report high feature variability [21], whereas studies using greyscale mapping yielded more reproducible results [19,24]. Furthermore, when using a fixed bin size scheme with the lower bound on zero [26], we found increased variability in the absence of exposures with stable plateaus (S7 Fig). However, neither feature variability (Fig 6) nor the occurrence of stable plateaus (Fig 7), were affected much by GLCM bit size.

We tested three types of standard image reconstruction algorithms, namely FBP, OSEM and TrueX at standard (4 mm) and high (2 mm) resolutions, and a seventh TrueX reconstruction at ultra-high resolution (1.6 mm). Through these systematic measurements, we found for the SR dataset a large subset of features that proved to be stable at an exposure of approximately 200 kBq/ml*s. Excursions to high resolution yielded stable features only at 10-fold higher exposure, but low-resolution reconstructions required exposures falling in the range encountered in clinical routine. However, we also found 14 feature-reconstruction combinations with stability over the entire exposure range. On the other hand, angular second moment, variance, cluster shade, and prominence calculated from normalized GLCMs all showed exposure variability of up to 2.5 magnitudes when used in combination with HR OSEM and HR PSF reconstructions (Fig 6B). In a clinical setup, there is always a tradeoff between diagnostic utility of features and variability brought by the use of these features.

Fig 7 shows domains of stability occurring for most of the eleven features at standard resolution without a requirement for high or excessive exposure times. This finding is a matter of practical interest in clinical routine, and demonstrates the usefulness of feature values at our standard resolution. Because our objective is to obtain the best possible image texture information without increasing exposure, the chosen resolution has to match the clinical question. First and foremost, a chosen image resolution must support the visualization of a lesion’s true texture to justify the use of texture analysis over a simple tracer uptake analysis. Next, the resolution must provide for sufficient voxels in a lesion to satisfy the statistical nature of Haralick features; small lesions might simply not have enough neighboring voxels for a concise feature description. Our results are not promising for the detection of features in lesions measuring < 12 mm, i.e. lesions with less than three SR-voxels in any direction.

Using a daily QA phantom for estimating correct acquisition times facilitates confirming that a given PET/CT system together with the chosen reconstruction method suffice to yield useful images within a practical time and tracer dose limits. The required acquisition duration will vary in conjunction with sensitivity of different PET/CT systems. All else being equal, a system with tenfold lower sensitivity than ours would require ten times longer scan durations for the same task. Furthermore, differing definitions of how GLCMs are calculated prevent the use of simple look-up tables for exposures, and thus mandate individual phantom measurements. With this work, we present an easily repeatable procedure to identify stable exposure domains, where image noise will not substantially add to textural feature values. These domains should be seen as minimal requirements for comparable radiomics in a multicenter setting. In addition, to quantify a heterogeneous texture, one must first be able to identify truly homogeneous areas, where any seeming texture is actually image noise.

However, there are certain caveats to using a perfectly homogeneous phantom. First, as the covariance matrix in OSEM and PSF based image reconstructions depends on the very object being imaged [36], making detailed predictions about the exposure dependence of texture feature values becomes difficult in heterogeneous clinical objects. Additionally, the non-Gaussian, object-dependent PSFs of said image reconstruction algorithms [50] violate the scale space axioms, stating that the imaging process should create no new image structures [51]. Particularly, PSF based reconstructions violate the axiom of non-creation and non-enhancement of local extrema and, and as we have shown previously [50], the scale invariance axiom. Furthermore, due to regularization processes in these algorithms, the membership of images created with PSF-based reconstructions in the Lie-group is disputable over the relevant scales. Therefore, acquisition durations determined for OSEM and PSF reconstructions depict only a general minimum time domain where proper texture analysis becomes possible. However, some of these problems could be compounded for inhomogeneous phantom measurements with real texture. Nevertheless, it is important to keep in mind that textural features calculated with nearest neighborhood operations by definition represent properties found at the highest spatial frequencies of the image. The highest spatial frequencies are precisely those most affected by the actual modulation transfer function (MTF) of the tomograph, by image resolution in terms of sampling, by the SNR, and by non-Gaussian distributions of image noise. In fact, textural features calculated from differing sampling densities characterize different features found at different spatial frequencies. Such features are scale-variant. This state of affairs calls for the use of scale-invariant phantoms, because any scale-variant phantom will produce differing results for differing spatial resolutions. As such, it can be said that a homogenous phantom fits all spatial resolutions and reconstructions.

As a second caveat, the recovery of decays depends also on the depth of the VOI within the body and the size of the body [31]. Using the daily QA phantom, with its inner diameter of 200 mm, can thus lead to underestimated acquisition durations in the case of deep lesions with higher attenuation and scatter. Therefore, calculating the correct exposure for feature detection must consider the lesion depth and the applied scatter and attenuation correction. Our work used the deepest possible location for the chosen VOI; shallower locations can only bring an earlier onset of stable plateaus. Finally, our present use of the long-lived 68-Ge-radoinuclide in the phantom does not emulate the decreasing signal to noise ratio occurring with the decay of short-lived clinical PET tracers. However, this last pitfall is easily correctable by using the expected (integral of) total decays instead of the decay-corrected activity concentration of the nuclide for dynamic exposure calculations. Nevertheless, for the case of [68-Ga]-PSMA-11 imaging, with its 68 min physical half-life, correct exposures may be unobtainable in low uptake lesions. When using different nuclides, their branching ratios must be considered to calculate exposure values. Because positron trajectories are de-correlated from the MTF, differences in positron range are irrelevant between isotopes. Additionally, the transconvolution method allows for recasting images acquired with different positron emitters [52].

Our findings stand in contrast to those of other works [28,29], where feature values were deemed noise-invariant. Besides clearly showing emergence of texture at low exposures (Fig 1A and 1B), our data shows noise-dependence of texture. Because texture stability occurs at sufficiently high exposures, variability is easily overlooked in studies with long acquisitions such as performed in [28,29]. Furthermore, features calculated from neighboring voxels (d = 1voxel) describe texture exactly at the Nyquist frequency of an imaging system. Therefore, having a PSF with an effective FWHM of three to four times the voxel size [29], neighboring voxels will—regardless of the underlying texture or noise—show similar intensities. In this case, texture values will appear exposure-invariant, or the variance occurs at lower exposures. This apparent feature stability emerges because the MTF of such an imaging system will effectively filter structures and noise alike. Using a different feature definition with a displacement d > 1voxel would be advisable in cases where image oversampling is desired.

Conclusion

Proper textural feature analysis is attainable only when respecting the limitations of the PET/CT system, and with appropriate adjustment of the acquisition duration. While differences in imaging properties, e.g. how the image is formed, can be overcome with transconvolution [10], the range of sensitivity of contemporary PET/CT systems mandates individual adjustment of exposure times. Otherwise, a comparative use of statistical texture features across different PET/CT devices or even differing reconstruction algorithms is destined to fail. In the context of multicenter studies, harmonized acquisition protocols should match spatial resolution and sampling, but should actually use non-standard acquisition durations. These non-standard acquisition durations require individual tailoring to a particular clinical site and question. In this respect, we present a methodology to perform measurements to arrive at minimum exposures required for robust feature value identification before even proceeding to clinical imaging. Not only will the subsequent clinical measurements then vary less, but also expected feature values can be estimated for a particular PET/CT site.

Choosing a sub-set of stable features together with an appropriate image reconstruction protocol is every bit as important as adjusting exposure to fit the expected domain of stability. Based on our findings, we recommend using different image reconstructions for feature analyses (FBP or OSEM) and for viewing (PSF). When using FBP, exposure must be high enough to avoid negative image values. We also suggest always using greyscale mapping, together with grey level invariant feature descriptions [23]. An additional benefit of adjusting exposure for optimal texture detection will be lower sample size requirement in multicenter studies and, as is becoming ever more important, for machine learning algorithms in medical imaging. Furthermore, harmonizing acquisition protocols as suggested above could be the key to bringing texture feature analysis into clinical routine. Naturally, our findings from a homogeneous phantom require confirmation in the clinic for a wide range of tumor entities and tissue types. Furthermore, texture metrics other than Haralick features exist also merit close investigation.

Supporting information

S1 Fig. Variability of first order statistics.

Mean activity concentration μc in the VOI as a function of exposure and grouped according to different reconstruction protocols. (a) SR acquisitions, (b) HR and UHR acquisitions. (c) Shown is the absolute range between the lowest and highest values in units of Bel found in first order statistics. Liliefor’s p varied up to 12 orders of magnitude and is not shown to scale.

https://doi.org/10.1371/journal.pone.0229560.s001

(PDF)

S2 Fig. Effect of GLCM greyscale quantization on magnitudes of textural feature variability.

(a) Cumulative histograms of absolute ranges between the lowest and highest values in units of Bel found in all texture features from differently quantized GLCMs. (b) Histograms of differences in feature range from differently quantized GLCMs, i.e. effect size.

https://doi.org/10.1371/journal.pone.0229560.s002

(PDF)

S3 Fig. Full range GLCMs.

Overlaid examples of five full range GLCMs calculated from six differently reconstructed datasets with five selected acquisition durations. GLCM maxima are normalized to unity and are displayed with a logarithmic color scale. Acquisition times are noted adjacent to each SR FBP GLCM in the respective colors. NPairs denotes the number of GLCM pairs analyzed for every color-coded matrix. Color scale in HR:FBP applies to all. Example GLCMs for HR:PSF are not shown, due to their similarity to the UHR:PSF measurements. Greyscales are truncated at 400.

https://doi.org/10.1371/journal.pone.0229560.s003

(PDF)

S4 Fig. Islands of stability.

Calculating islands of stability for the case of the inverse difference moment (IDM) feature according to the seven different reconstruction protocols, and for full range GLCMs. Plots depict the LOESS curves (colored solid lines) of the data (symbols) with their 99% confidence intervals (dotted lines). Black solid and dotted lines show the respective first derivatives of the Loess curves, with zero crossings at the intersections with the dashed horizontal line. Black crosses mark the islands of stability found for the IDM. The middle legend applies to all but the UHR PSF data.

https://doi.org/10.1371/journal.pone.0229560.s004

(PDF)

S5 Fig. Feature values calculated from full range GLCMs.

Exposure dependency of Haralick textural features according to different image resolution and reconstruction algorithm calculated from full range GLCMs with 256 grey levels. Measurement points are shown in the first graph but omitted for clarity in subsequent. Loess curves are shown without associated confidence intervals.

https://doi.org/10.1371/journal.pone.0229560.s005

(PDF)

S6 Fig. Variability and stability in features from full range GLCMs.

(a) Variability between the lowest and highest values found in the texture features across all examined exposures calculated form full range GLCMs. (b) Islands of stability in features calculated from full range GLCMs with median exposure as a blue line. *Missing values indicate absence of an island of stability.

https://doi.org/10.1371/journal.pone.0229560.s006

(PDF)

S7 Fig. Effect of restricted range GLCM quantizing.

Exposure dependency of all examined textural features according to different image resolution and reconstruction algorithm binned into a GLCM with 512 grey levels restricted from zero to a maximal intensity of 25 kBq/ml. Measurement points are shown in the first graph, but omitted subsequently. Loess curves are shown without confidence intervals.

https://doi.org/10.1371/journal.pone.0229560.s007

(PDF)

S1 Text. Supporting material and methods.

Full range and restricted range GLCM quantization.

https://doi.org/10.1371/journal.pone.0229560.s008

(PDF)

S2 Text. Supporting results.

Assessing exposure stability of image features when using full range GLCMs.

https://doi.org/10.1371/journal.pone.0229560.s009

(PDF)

Acknowledgments

Thank goes to the staff at the Department of Nuclear Medicine, Inselspital Bern, for assistance with the PET phantom measurements.

References

  1. 1. Amadasun M, King R (1989) Textural features corresponding to textural properties. IEEE Transactions on Systems, Man, and Cybernetics 19: 1264–1274.
  2. 2. Kirienko M, Cozzi L, Rossi A, Voulaz E, Antunovic L, Fogliata A, et al. (2018) Ability of FDG PET and CT radiomics features to differentiate between primary and metastatic lung lesions. European Journal of Nuclear Medicine and Molecular Imaging 45: 1649–1660. pmid:29623375
  3. 3. Nakajo M, Jinguji M, Nakajo M, Shinaji T, Nakabeppu Y, Fukukura Y, et al. (2017) Texture analysis of FDG PET/CT for differentiating between FDG-avid benign and metastatic adrenal tumors: efficacy of combining SUV and texture parameters. Abdominal Radiology 42: 2882–2889. pmid:28612161
  4. 4. Chicklore S, Goh V, Siddique M, Roy A, Marsden PK, Cook GJR (2013) Quantifying tumour heterogeneity in 18F-FDG PET/CT imaging by texture analysis. European Journal of Nuclear Medicine and Molecular Imaging 40: 133–140. pmid:23064544
  5. 5. Lee HS, Oh JS, Park YS, Jang SJ, Choi IS, Ryu J-S (2016) Differentiating the grades of thymic epithelial tumor malignancy using textural features of intratumoral heterogeneity via (18)F-FDG PET/CT. Annals of nuclear medicine 30: 309–319. pmid:26868139
  6. 6. Tixier F, Le Rest CC, Hatt M, Albarghach N, Pradier O, Metges J-P, et al. (2011) Intratumor Heterogeneity Characterized by Textural Features on Baseline 18F-FDG PET Images Predicts Response to Concomitant Radiochemotherapy in Esophageal Cancer. Journal of Nuclear Medicine 52: 369–378. pmid:21321270
  7. 7. Yu H, Caldwell C, Mah K, Poon I, Balogh J, MacKenzie R, et al. (2009) Automated Radiation Targeting in Head-and-Neck Cancer Using Region-Based Texture Analysis of PET and CT Images. International Journal of Radiation Oncology, Biology, Physics 75: 618–625. pmid:19683403
  8. 8. Haralick RM, Shanmugam K, Dinstein IH (1973) Textural Features for Image Classification. Systems, Man and Cybernetics, IEEE Transactions on SMC-3: 610–621.
  9. 9. Tesař L, Shimizu A, Smutek D, Kobatake H, Nawano S (2008) Medical image analysis of 3D CT images based on extension of Haralick texture features. Computerized Medical Imaging and Graphics 32: 513–520. pmid:18614335
  10. 10. Prenosil GA, Weitzel T, Hentschel M, Klaeser B, Krause T (2013) Transconvolution and the virtual positron emission tomograph–A new method for cross calibration in quantitative PET/CT imaging. Medical Physics 40: 15. pmid:23718608
  11. 11. Mansor S, Pfaehler E, Heijtel D, Lodge MA, Boellaard R, Yaqub M (2017) Impact of PET/CT system, reconstruction protocol, data analysis method, and repositioning on PET/CT precision: An experimental evaluation using an oncology and brain phantom. Medical Physics 44: 6413–6424. pmid:28994465
  12. 12. Fahey FH, Kinahan PE, Doot RK, Kocak M, Thurston H, Poussaint TY (2010) Variability in PET quantitation within a multicenter consortium. Medical Physics 37: 3660–3666. pmid:20831073
  13. 13. Boellaard R, Delgado-Bolton R, Oyen WJG, Giammarile F, Tatsch K, Eschner W, et al. (2015) FDG PET/CT: EANM procedure guidelines for tumour imaging: version 2.0. European Journal of Nuclear Medicine and Molecular Imaging 42: 328–354. pmid:25452219
  14. 14. Aide N, Lasnon C, Veit-Haibach P, Sera T, Sattler B, Boellaard R (2017) EANM/EARL harmonization strategies in PET quantification: from daily practice to multicentre oncological studies. European journal of nuclear medicine and molecular imaging 44: 17–31. pmid:28623376
  15. 15. Schmidtlein CR, Beattie BJ, Bailey DL, Akhurst TJ, Wang W, Gönen M, et al. (2010) Using an external gating signal to estimate noise in PET with an emphasis on tracer avid tumors. Physics in Medicine and Biology 55: 6299–6326. pmid:20924132
  16. 16. Doot RK, Scheuermann JS, Christian PE, Karp JS, Kinahan PE (2010) Instrumentation factors affecting variance and bias of quantifying tracer uptake with PET/CT. Medical physics 37: 6035–6046. pmid:21158315
  17. 17. Siman W, Mawlawi OR, Mikell JK, Mourtada F, Kappadath SC (2016) Effects of image noise, respiratory motion, and motion compensation on 3D activity quantification in count-limited PET images. Physics in Medicine and Biology 62: 448–464. pmid:28000618
  18. 18. Molina-Duran F, Dinter D, Schoenahl F, Schoenberg SO, Glatting G (2014) Dependence of image quality on acquisition time for the PET/CT Biograph mCT. Zeitschrift fur medizinische Physik 24: 73–79. pmid:23561551
  19. 19. Tixier F, Hatt M, Le Rest CC, Le Pogam A, Corcos L, Visvikis D (2012) Reproducibility of Tumor Uptake Heterogeneity Characterization Through Textural Feature Analysis in 18F-FDG PET. Journal of Nuclear Medicine 53: 693–700. pmid:22454484
  20. 20. Zwanenburg A (2019) Radiomics in nuclear medicine: robustness, reproducibility, standardization, and how to avoid data analysis traps and replication crisis. European Journal of Nuclear Medicine and Molecular Imaging. pmid:31240330
  21. 21. Galavis PE, Hollensen C, Jallow N, Paliwal B, Jeraj R (2010) Variability of textural features in FDG PET images due to different acquisition modes and reconstruction parameters. Acta Oncol 49: 1012–1016. pmid:20831489
  22. 22. Shafiq-ul-Hassan M, Latifi K, Zhang G, Ullah G, Gillies R, Moros E (2018) Voxel size and gray level normalization of CT radiomic features in lung cancer. Scientific Reports 8: 10545. pmid:30002441
  23. 23. Löfstedt T, Brynolfsson P, Asklund T, Nyholm T, Garpebring A (2019) Gray-level invariant Haralick texture features. PLOS ONE 14: e0212110. pmid:30794577
  24. 24. Desseroit M-C, Tixier F, Weber WA, Siegel BA, Cheze Le Rest C, Visvikis D, et al. (2017) Reliability of PET/CT Shape and Heterogeneity Features in Functional and Morphologic Components of Non–Small Cell Lung Cancer Tumors: A Repeatability Analysis in a Prospective Multicenter Cohort. Journal of Nuclear Medicine 58: 406–411. pmid:27765856
  25. 25. Bailly C, Bodet-Milin C, Couespel S, Necib H, Kraeber-Bodéré F, Ansquer C, et al. (2016) Revisiting the Robustness of PET-Based Textural Features in the Context of Multi-Centric Trials. PLOS ONE 11: e0159984. pmid:27467882
  26. 26. Presotto L, Bettinardi V, De Bernardi E, Belli ML, Cattaneo GM, Broggi S, et al. (2018) PET textural features stability and pattern discrimination power for radiomics analysis: An "ad-hoc" phantoms study. Phys Med 50: 66–74. pmid:29891096
  27. 27. Pfaehler E, Beukinga RJ, de Jong JR, Slart R, Slump CH, Dierckx R, et al. (2019) Repeatability of (18) F-FDG PET radiomic features: A phantom study to explore sensitivity to image reconstruction settings, noise, and delineation method. Medical physics 46: 665–678. pmid:30506687
  28. 28. Grootjans W, Tixier F, van der Vos CS, Vriens D, Le Rest CC, Bussink J, et al. (2016) The Impact of Optimal Respiratory Gating and Image Noise on Evaluation of Intratumor Heterogeneity on 18F-FDG PET Imaging of Lung Cancer. Journal of Nuclear Medicine 57: 1692–1698. pmid:27283931
  29. 29. Shiri I, Rahmim A, Ghaffarian P, Geramifar P, Abdollahi H, Bitarafan-Rajabi A (2017) The impact of image reconstruction settings on 18F-FDG PET radiomic features: multi-scanner phantom and patient studies. European Radiology 27: 4498–4509. pmid:28567548
  30. 30. Reuze S, Schernberg A, Orlhac F, Sun R, Chargari C, Dercle L, et al. (2018) Radiomics in Nuclear Medicine Applied to Radiation Therapy: Methods, Pitfalls, and Challenges. International journal of radiation oncology, biology, physics 102: 1117–1142. pmid:30064704
  31. 31. Nyflot MJ, Yang F, Byrd D, Bowen SR, Sandison GA, Kinahan PE (2015) Quantitative radiomics: impact of stochastic effects on textural feature analysis implies the need for standards. Journal of medical imaging (Bellingham, Wash) 2: 041002. pmid:26251842
  32. 32. Schottky W (1918) Über spontane Stromschwankungen in verschiedenen Elektrizitätsleitern. Annalen der Physik 362: 541–567.
  33. 33. Teymurazyan A, Riauka T, Jans H-S, Robinson D (2013) Properties of Noise in Positron Emission Tomography Images Reconstructed with Filtered-Backprojection and Row-Action Maximum Likelihood Algorithm. Journal of Digital Imaging 26: 447–456. pmid:22850933
  34. 34. Shepp LA, Vardi Y (1982) Maximum Likelihood Reconstruction for Emission Tomography. Medical Imaging, IEEE Transactions on 1: 113–122.
  35. 35. Vardi Y, Shepp LA, Kaufman L (1985) A Statistical Model for Positron Emission Tomography. Journal of the American Statistical Association 80: 8–20.
  36. 36. Barrett HH, Wilson DW, Tsui BM (1994) Noise properties of the EM algorithm: I. Theory. Physics in medicine and biology 39: 833–846. pmid:15552088
  37. 37. Wilson DW, Tsui BM, Barrett HH (1994) Noise properties of the EM algorithm: II. Monte Carlo simulations. Physics in medicine and biology 39: 847–871. pmid:15552089
  38. 38. Riederer SJ, Pelc NJ, Chesler DA (1978) The noise power spectrum in computed X-ray tomography. Phys Med Biol 23: 446–454. pmid:674361
  39. 39. Qi J (2003) A unified noise analysis for iterative image estimation. Physics in medicine and biology 48: 3505–3519. pmid:14653559
  40. 40. Arman R, Jing T (2013) Noise propagation in resolution modeled PET imaging and its impact on detectability. Physics in Medicine and Biology 58: 6945. pmid:24029682
  41. 41. Rahmim A, Qi J, Sossi V (2013) Resolution modeling in PET imaging: Theory, practice, benefits, and pitfalls. Medical Physics 40: 064301. pmid:23718620
  42. 42. Panin VY, Kehren F, Michel C, Casey M (2006) Fully 3-D PET reconstruction with system matrix derived from point source measurements. IEEE transactions on medical imaging 25: 907–921. pmid:16827491
  43. 43. Van-Roy P, Haridi S (2004) Concepts, Techniques, and Models of Computer Programming: Prentice-Hall. ISBN: 9780262220699
  44. 44. Lilliefors HW (1967) On the Kolmogorov-Smirnov Test for Normality with Mean and Variance Unknown. Journal of the American Statistical Association 62: 399–402.
  45. 45. Albregtsen F (2008) Statistical Texture Measures Computed from Gray Level Coocurrence Matrices. In: Image Processing Laboratory DoI, University of Oslo, editor: University of Oslo. pp. 13.
  46. 46. Leijenaar RTH, Nalbantov G, Carvalho S, van Elmpt WJC, Troost EGC, Boellaard R, et al. (2015) The effect of SUV discretization in quantitative FDG-PET Radiomics: the need for standardized methodology in tumor texture analysis. Scientific Reports 5: 11075. pmid:26242464
  47. 47. Krarup MMK, Nygård L, Vogelius IR, Andersen FL, Cook G, Goh V, et al. (2020) Heterogeneity in tumours: Validating the use of radiomic features on 18F-FDG PET/CT scans of lung cancer patients as a prognostic tool. Radiotherapy and Oncology 144: 72–78. pmid:31733491
  48. 48. Schwartz J, Grkovski M, Rimner A, Schöder H, Zanzonico PB, Carlin SD. et al. (2017) Pharmacokinetic Analysis of Dynamic 18F-Fluoromisonidazole PET Data in Non–Small Cell Lung Cancer. Journal of Nuclear Medicine 58: 911–919. pmid:28232611
  49. 49. Bettinardi V, Presotto L, Rapisarda E, Picchio M, Gianolli L, Gilardi MC (2011) Physical performance of the new hybrid PETCT Discovery-690. Med Phys 38: 5394–5411. pmid:21992359
  50. 50. Prenosil GA, Klaeser B, Hentschel M, Fürstner M, Berndt M, Krause T, et al. (2016) Isotope independent determination of PET/CT modulation transfer functions from phantom measurements on spheres. Medical Physics 43: 5767–5778. pmid:27782715
  51. 51. Lindeberg T (2011) Generalized Gaussian Scale-Space Axiomatics Comprising Linear Scale-Space, Affine Scale-Space and Spatio-Temporal Scale-Space. Journal of Mathematical Imaging and Vision 40: 36–81.
  52. 52. Prenosil GA, Hentschel M, Furstner M, Krause T, Weitzel T, Klaeser B (2017) Technical Note: Transconvolution based equalization of positron energy effects for the use of 68 Ge/68 Ga phantoms in determining 18 F PET recovery. Med Phys 44: 3761–3766. pmid:28494090