Physics-Informed Neural Network with Fourier Features for Radiation Transport in Heterogeneous Media

Abstract Typical machine learning (ML) methods are difficult to apply to radiation transport due to the large computational cost associated with simulating problems to create training data. Physics-informed Neural Networks (PiNNs) are a ML method that train a neural network with the residual of a governing equation as the loss function. This allows PiNNs to be trained in a low-data regime in the absence of (experimental or synthetic) data. PiNNs also are trained on points sampled within the phase-space volume of the problem, which means they are not required to be evaluated on a mesh, providing a distinct advantage in solving the linear Boltzmann transport equation, which is difficult to discretize. We have applied PiNNs to solve the streaming and interaction terms of the linear Boltzmann transport equation to create an accurate ML model that is wrapped inside a traditional source iteration process. We present an application of Fourier Features to PiNNs that yields good performance on heterogeneous problems. We also introduce a sampling method based on heuristics that improves the performance of PiNN simulations. The results are presented in a suite of one-dimensional radiation transport problems where PiNNs show very good agreement when compared to fine-mesh answers from traditional discretization techniques.


I. INTRODUCTION
Often, machine learning (ML) processes are data driven and rely on the acquisition of a large amount of data in order to be trained. ML has been applied to many fields in science and engineering, 1,2 where these training data, be they empirical measurements or synthetic results generated from simulations, can be costly to obtain in terms of resources (human time and hardware/software costs). In addition, data-driven ML processes may not yet be fully generalizable or explainable, and purely data-driven methods may not extrapolate well due to observational biases. 3 Therefore, there is a need for incorporating governing laws [e.g., partial differential equations (PDEs)] into the learning process in order to increase the robustness of ML techniques in low-data regimes and to improve their explainability. When concerned with the transport of radiation (neutrons, photons, etc.) through a background medium, one has to discretize a fairly highdimensional phase-space in order to track particles with respect to their position, energy, and direction of motion.
In three-dimensional geometries, this results in a sixdimensional phase-space, making the discretization of the associated linear Boltzmann transport equation a challenging task, even on modern computer architectures. 4 This potentially exacerbates the low-data regime issues mentioned above, as numerical solutions to the transport equation are costly to obtain.
In this work, we present the application of physicsinformed neural networks (PiNNs) to the radiation transport equation. PiNNs can be categorized as a deep-learning technique based on (deep) feedforward neural networks.
Typically, the input layer of a PiNN consists of one neuron per independent phase-space variable, followed by several hidden layers with a significant number of neurons per layer (which makes this a deep-learning approach), and finally an output layer consisting of one neuron (for scalar PDEs) or several neutrons (for vector PDEs). At the heart of a PiNN feedforward neural network is the mechanism by which the parameters (weights and biases) of the network are obtained. This is directly related to the definition of the neural network's loss function. The loss function incorporates (1) the governing laws themselves (i.e., the PDE's residual in the phase-space volume), thus making this a physics-informed approach; (2) boundary conditions; (3) initial conditions (for time-dependent problems); and (4) measurements (if any).
PiNNs were first proposed in Ref. 5 and utilize the automatic differentiation capabilities of the neural network to evaluate differential operators occurring in the PDE, hence enabling the construction of the loss function (i.e., residual) associated with the PDE. It must be noted that the differentiation is, therefore, taken with respect to the input layer variables (e.g., position and time, as needed) to build the residual of the PDE at selected phase-space points. The standard backpropagation algorithm is employed for this task. In order to optimize the network's parameters, automatic differentiation is also employed with respect to the network's parameters (this is the more commonly understood use of backpropagation in neural networks).
Parameter optimization (i.e., the training of the PiNN) relies on standard optimization algorithms in ML, specifically, Adam optimization 6 and limited-memory BFGS (L-BFGS), a quasi-Newton gradient-based optimization algorithm. 7 A typical PiNN architecture as well as elements of the process behind PiNNs are shown in Fig. 1. In this schematic, the neural network consists of two input neurons [a two-dimensional (2-D) phase-space described by independent variables x and μ]; this is followed by n hidden layers, each with m neurons, and an output layer for the scalar variable ψðx; μÞ that is the solution of the PDE F ðψðx; μÞÞ ¼ q. The residual of the PDE at sampled points in the phase-space volume is built, as well as the boundary residual. Both residuals are then used in the definition of a loss function whose minimization process trains the parameters of the neural network.
Automatic differentiation is used at two instants: first, to evaluate the residual of the PDE (as governing laws often contain spatial derivatives of the solution variables) and second, to optimize the weights and biases of the network. Given that the PDE residual is evaluated at sampled points in the phase-space volume, a PiNN solution does not require meshing of the geometry, and hence, PiNN techniques to solve PDEs are often thought of as a family of meshless methods.
PiNNs have already been applied to a wide variety of disciplines in science and engineering; see, for instance, the recent review paper 3 and the references therein. Despite early successes in using PiNNs to solve governing laws using ML, several limitations currently exist. For example, a solid mathematical explanation of PiNNs is still lacking, which precludes understanding why PiNNs can oftentimes fail. In addition, the large Fig. 1. PiNN example: The neural network consists of two input neurons (in the case of a 2-D phase-space), followed by n hidden layers, each with m neurons, and an output layer for the scalar variable, solution of the PDE F ðψðx; μÞÞ ¼ q. The residual of the PDE at sampled points in the phase-space volume is built, as well as the boundary residual. Both residuals are then in the definition of a loss function whose minimization process trains the parameters of the neural network.  8, it is noted that PiNNs can struggle to converge when the solution exhibits high frequencies or multiscale features (or simply highly heterogeneous material properties, per our own observations); the authors showed that fully connected PiNNs converge to Gaussian processes in the infinite width limit for linear PDEs. The convergence rate of the total training error of a PiNN model was analyzed in terms of the spectrum of its neural tangent kernel.
In Ref. 9, the authors employ Fourier Feature networks that use a simple Fourier Feature mapping to increase the ability of fully connected neural networks (fcNNs) to learn high-frequency functions (this approach was originally proposed in Ref. 10). For advectiondominated problems, additional techniques have been proposed to improve the accuracy of PiNNs. In Ref. 11 the authors argue that the parameter optimization that takes place in the PiNN learning process does not respect the space-time causality of physical systems where information propagates at a finite speed; to respect the physical causality during the learning process, they proposed to reformulate the definition of the loss function so that learning (and thus parameter optimization) accounts for physical causality.
In another study, the authors of Ref. 12 also note the failure of PiNNs to converge for "complicated" PDEs. They were able to overcome some of the challenges by developing adaptive sampling strategies used in training the PiNNs. They hypothesized that training of PINNs relies on a successful "propagation" of solution from initial and/or boundary condition points to interior points, and they subsequently proposed evolutionary sampling algorithms to incrementally add sampling points in the location of high residuals.
In Ref. 13, the governing laws are rewritten in a Lagrangian framework, so that the PiNNs now include two distinct branches: one that learns the lowdimensional characteristics curve and one that solves for the state variables along the characteristics. This approach, coined Lagrangian PiNNs, was shown to outperform the traditional PiNNs in the Eulerian framework for a wide class of convection-diffusion problems.
Finally, another approach developed to combat the lack of high-accuracy solutions of PiNNs consists of adding a discriminator network in addition to the PiNN. In this adversarial approach, called competitive PiNNs, the discriminator is trained (rewarded) for predicting mistakes the PiNN commits.
The remainder of this paper is as follows. In Sec. II, we review the radiation transport equation and provide some details related to the standard scattering source iteration (SI) solution process that is also employed within our PiNN definitions. In Sec. III, we provide details on the PiNNs (phase-space inputs, definition of loss function, and handling of scattering terms). We pay particular attention to improvements made to handle heterogeneous material via a combination of pass-through and Fourier Feature neurons after the input layer. Results are provided in Sec. IV for various one-dimensional (1-D) radiation transport benchmarks. Conclusions and outlook are proposed in Sec. V.

II. RADIATION TRANSPORT GOVERNING LAWS
The linear Boltzmann equation for neutral particles, in steady state, with isotropic scattering and external sources and single-speed formalism, is given by In 1-D slab geometry, the streaming operator Ω �Ñ devolves to μ q qx , with μ ¼Ω �ẽ x . For classical discretization techniques, Eq. (1) is discretized in space and angle. We recall briefly these discretization approaches here, although a PiNN solution process is meshless and does not require the notion of grid or meshes. For traditional techniques, an angular quadrature is often used to discretize the angular variable and to carry out angular integrations: supplemented with boundary conditions: with the scalar flux being evaluated as The angular quadrature is given by the direction-weight pairs ðΩ d ; w d Þ with ð1 � d � N Ω ), and the approach is referred to in the literature as a discrete ordinates angular discretization. 14 Note that Ψ d ðrÞ ; Ψðr;Ω d Þ. For spatial discretization, it is standard to use finite differencing (such as the diamond scheme, Ref. 15) or discontinuous finite elements. 16,17 We refer the reader to these references for additional spatial discretization details. In what follows, we will keep using the spatially undiscretized expressions of Eq. (3) for brevity.
With regard to solution techniques, both the traditionally discretized transport equation and its PiNN equivalent will employ a standard Source Iteration (SI) approach. In SI, the scattering source is lagged at the previous iteration ,; and the solution process is as follows: (1) Given old values for the scalar flux, obtain new angular fluxes: (2) Update the scalar flux:

III. PiNNs: DEFINITIONS AND ADVANCED FEATURES
In this section, we detail the specifics of our application of PiNNs to radiation transport. In Sec. III.A, we describe our implementation of PiNN for each angular direction and how the angular flux estimates for each PiNN are used to update the scalar flux through SIs. In Sec. III.B, we give a brief overview of fully connected neural network architecture. In Sec. III.C, we detail our specific application of Fourier Features to improve PiNN performance in heterogeneous problems. In Sec. III.D, we give a description of the training of neural networks as well as how PiNNs differ from traditional supervised neural networks. In Sec. III.E, we explain how PiNNs leverage automatic differentiation to calculate the partial derivatives in the governing equation (here, the radiation transport equation). Finally, in Sec. III.F, we explain our sampling methods that lead to better training performance in heterogeneous problems. These specific methods allow for PiNNs to be applied to the more difficult heterogeneous problems that a vanilla implementation of PiNNs struggles to solve effectively, if at all.

III.A. Source Iteration and PiNNs
As is described in Sec. II, SI is often employed to solve radiation transport problems. In order to easily solve for the scalar flux, a PiNN for each direction in the angular quadrature is devised. Specifically, the PiNN is substituted for the inversion of the left-hand side operator of Eq. (4a); this is shown in Eq. (5): where the left-hand side of the equation represents the output of the PiNN trained while being driven by the current total (i.e., extraneous + scattering) source evaluated using the lagged scalar flux Φ ð,Þ . Once all angular fluxes Ψ ð,þ1Þ d have been obtained by training N Ω different PiNNs, we can update the scalar flux by using Eq. (4b) as before. The scalar flux is then used to update the right-PHYSICS-INFORMED NEURAL NETWORK WITH FOURIER FEATURES · HUHN et al. 2487 hand side of the governing equation, as traditionally done in SIs. Hence, in our scheme, the SIs are wrapped around the classical transport sweeps; however, PiNNs are used and trained in order to obtain those angular fluxes (one PiNN per angular direction). Note that between each update of the angular flux, the PiNN is not initialized again but instead uses the final state of the previous SI. Because of this, both the SI error and training error are reduced together instead of needing to fully retrain each PiNN at each SI. This also means that the number of epochs required to fully train each PiNN decreases with each SI, as will be illustrated in our numerical results.

III.B. Neural Network Architecture
As shown in Fig. 1, fcNNs are composed of an input layer, hidden layers, and an output layer. A fcNN can have any number of inputs and outputs, and in general, both can be normalized or scaled. There will be one neuron for each input and output, and each hidden layer can contain any number of neurons. Between hidden layers l À 1 and l, there will be a set of weights W ðlÞ 2 R ðm lÀ 1 �m l Þ , where m l is the number of neurons in layer l. The weights are used to generate a linear combination of the previous layer's output. This linear combination will form the input for the next layer. In addition, each neuron in layer l also has a bias, i.e., b ðlÞ 2 R m l . Nonlinear activation functions a (typically, sigmoid, tanh, and ReLU) transform the input of each neuron. Hence, the output of a given layer in a network from the output of the previous layer is given in Eq. (6): with z ðlÞ 2 R m l . Given a total of n hidden layers, the fully connected network of Fig. 1 can be mathematically described as with dim i and dim o the dimension of the input and output layers (in Fig. 1, these are 2 and 1), and the input/output layer functions Once trained (i.e., all weights and biases are optimized), the computation of this expression can be done very quickly, even for large networks. The main issue in utilizing a fcNN comes in the determination of the weights and biases, which form the set of trainable variables, hereinafter denoted as θ. This determination is achieved through training, which is explained in Sec. III.D.

III.C. Fourier Features
The benefit of Fourier Features to improve PiNNs' representation of high-frequency functions or problems with multiscale features is presented in Ref. 9. Drastically heterogeneous problems are generally difficult for PiNNs to represent, and Fourier Features promise to alleviate this issue. The general architecture for a neural network with Fourier Features is shown in Fig. 2, where Fourier Features project the neural network input by applying sines and cosines of various frequencies as shown in Eq. (7): where ω i is the i'th randomly sampled frequency sampled from the Gaussian distribution N ð0; σ 2 Þ; where σ is a hyperparameter of the network. In some implementations of Fourier Features, ω i are trained alongside the weights and biases of the network; however, in this paper they will be kept fixed at the sampled values.
A further modification to the Fourier Feature algorithm proposed is the addition of bypass neurons to the Fourier Feature layer of the network. These neurons will simply pass the input directly as is, instead of projecting in sine or cosine functions. The addition of bypass neurons leads to 2 � N FF þ N BN total neurons in the Fourier Feature layer, where N FF is the number of Fourier Feature pairs and N BN is the number of bypass neurons.

III.D. Neural Network Training
Neural networks are trained by optimizers that adjust the weights and biases in order to find an optimal version of the network. This optimal version is determined by the minimization of a loss function. In supervised fcNNs, the loss function is classically an error metric compared to a known or desired solution. However, for PiNNs, the loss function instead is based on the residual of the PDE under consideration. The method by which PiNNs are able to evaluate this residual is given later, in Sec. III.E. For now, it suffices to consider that the PiNN loss function is given by the sum of an interior residual and a boundary residual, as shown in Eq.

III.E. PDE Residual Calculation
PiNNs require a method for calculating the residual of the PDEs under consideration. To do so, PiNNs leverage automatic differentiation to calculate the partial derivatives found in the expression of the PDE. Automatic differentiation is also used in the training of traditional data-driven feedforward networks. This section will give a brief view of automatic differentiation; see Ref. 18 for additional details. The specific type of automatic differentiation used is reverse accumulation, which is composed of two steps: forward pass and backpropagation. The forward pass is simply an evaluation of the neural network where values for each neuron are saved to be used in the second step. When evaluating the forward pass for PiNNs, the values of the input must be saved as well. Once all of the relevant values are saved, it is then possible to perform a backpropagation stage.
Backpropagation allows the derivative of the output to be calculated with respect to any value within the network by leveraging the chain rule. While it is common to see backpropagation used with respect to parameters, we describe below how the derivative of the network's output (the angular flux in direction d in our case) with respect to input variables is computed. For example, to calculate qψ qx 1 from Fig. 3, the first application of the chain rule is shown in Eq. (10):

PHYSICS-INFORMED NEURAL NETWORK WITH FOURIER FEATURES · HUHN et al. 2489
This equation can be further expanded using the chain rule again, yielding where z i is the value of neuron i before the activation function is applied. It is then possible to analytically calculate qa ð1Þ i qz i for a given activation function a i . It can then be very easily found that the other two terms simply are the weights W ðlÞ for their respective layer l. This yields the final formula for qψ qx 1 , shown in Eq. (12), where the total contribution of each individual contribution is summed: Here, this approach is used in PiNNs in order to calculate the term containing qΨ qx in the formula for R int . Then, it is simple to sum the contribution from each term appearing in the PDE, yielding a complete method to compute the residual of said PDE: where Ψ d;i and Φ i are the angular flux in direction d and the scalar flux, both evaluated at internal point x i Ψ d;i is the output of the PiNN for direction d. Next, for the complete evaluation of the loss LðθÞ, the boundary residual R BC is computed as follows: where Ψ BC;d;j and Ψ d; j are the boundary condition and the angular flux (in direction d) evaluated at the boundary point x j . Again, Ψ d;j is the output of the PiNN for direction d. The sampling of the internal and boundary points is described in the next section.

III.F. Point Sampling
Point sampling is necessary to evaluate the loss function, and the sampling points are divided into boundary points and internal points. In this section, the different sampling approaches used in this paper are discussed. Boundary points are simply divided evenly between the two boundaries (left/ right) of the 1-D problems considered in this paper. They will not be discussed further. The simplest sampling approach for internal points is a random uniform sampling of a given number of points N i inside each material zone i in the problem domain. This method has advantages in its simplicity, but it usually does not perform as well at material interfaces, especially for highly heterogeneous configurations. As a first attempt to improve the performance of PiNNs, sampling is done such that there is a sufficient density of training points, given the mean-free-path thickness of each material zone i.
Equation (15) shows the formula for N i;mfp , the number of points in material zone i that ensures a density in each mean free path of material: where μ min is the smallest positive cosine of the angular directions, h i is the width of zone i, and α is a parameter that determines how many points will be sampled per mean free path. This sampling method ensures that each part of the domain is sampled enough to resolve any material boundaries or other heterogeneities. However, in problems where there is a large difference in material properties, there can still be difficulties in converging the solution well for all regions. We propose a heuristic, whereby the number of samples per region obtained from mean-free-path considerations is modified as follows. The number of points sampled is further increased in certain regions according to their influence, denoted γ i , on the loss function. The influence γ i is calculated with Eq. (16): where ϵ is a parameter that ensures that the number of points sampled will not become too large in spatially thin, optically thin, or void zones of the problem. The number of points to be sampled in each zone can be calculated using the zonal influences, as is shown in Eq.

IV. RESULTS
In this section, we present numerical results for our PiNN implementation applied to a set of five radiation benchmark problems taken from Ref. 19. The traditional Reed problem 20 is Problem 5 in our suite of test cases. All problems are evaluated with an S 8 quadrature except for Problem 3 and Problem 4', which are only evaluated with an S 2 quadrature (we used a lower-order angular quadrature in order to illustrate the effects of Fourier Features on the angular flux in Problem 3; the same conclusions hold, of course, for an S 8 angular quadrature as well). All problems use vacuum boundary conditions.
All of the PiNNs used to solve these problems employ the same architecture. They all start with a Fourier Feature projection layer followed by five fully connected layers. Each of the fully connected layers contains 64 neurons, and the Fourier Feature layers have 10 bypass neurons, 64 sine projection neurons, and 64 cosine projection neurons, for a total of 138 neurons in the Fourier layer. Each of the fully connected layers uses the swish activation function, and the inputs are normalized between zero and one before the Fourier Feature projection layer. Each PiNN was also trained using a combination of the Adam and LBFGS optimizers.
For each problem, the extent of each material spatial zone and the material properties of that zone (absorption and isotropic scattering macroscopic cross sections) will be given in a table. The flux solutions of the PiNNs are compared against a fine-mesh finite difference (FD) solution (here, we employed diamond-differencing with a fine enough mesh).
In terms of computing time, the FD traditional solvers yield the answer in a mere few seconds at most; training a PiNN on a single CPU on a modern desktop computer takes on the order of 60 min; with GPU acceleration, the training time is on the order of 20 min. These sample timing results were produced using the free version of the Google Colaboratory run-time environment for consistency in our comparisons. We note that the large factor in compute times between neural network methods and traditional solvers is consistent with observations reported elsewhere; see, for instance, Ref. 21.

IV.A. Problem 1
Problem 1 is a homogeneous slab containing a pure absorber. The extent of each spatial zone and the material properties are given in Table I. The results of the PiNNs compared to the results of a FD solution are shown in Fig. 4, where we observe excellent agreement between the two solutions. Homogeneous problems like this one are fairly easy to converge (even without using Fourier Features). The remainder of the test cases includes problems with more variations in the solution where the benefit of Fourier Features will become obvious.

IV.B. Problem 2
This test case includes the presence of scattering and is still a homogeneous material; however, the source is present in only one of the two material zones. The extent of each spatial zone and the material properties are given in Table II. The presence of scattering requires SIs to solve the problem. The results of the PiNNs compared to the results of a FD solution are shown in Fig. 5. There is good agreement between these two solutions. This problem proves that our approach of wrapping the SI around the PiNN solution is appropriate.
In Fig. 6a, we show the SI convergence (measured as the successive differences in scalar flux estimates). In Fig. 6b, we show a plot of the loss value versus epochs while training one of the neural networks (for one of the directions in the S N quadrature); in this graph, the sharp increase in the loss corresponds to a new scattering source provided from the SI process; between these SI updates, the loss decreases rapidly. Because the loss-versus-epochs graph (Fig. 6b) shows the loss for only one of the PiNNs, the loss error displayed does not directly correlate to the scalar flux error of Fig. 6a. As SIs continue, the number of epochs required for this error to decrease to its previous value (before the scattering source update) decreases until the scattering source converges.

IV.C. Problem 3
This problem and Problem 4 are both heterogeneous problems that will be employed to demonstrate the strengths and weaknesses of our neural network implementation. The extent of each spatial zone and the material properties are given in Table III.
Because of how heterogeneous this problem is, it is a good test case to illustrate the strength of the Fourier Features approach used in our neural network. The scalar flux solution for the PiNN with Fourier Features is shown in Fig. 7a. This can be compared to the scalar flux solution obtained without Fourier Features, shown in Fig. 8a. This discrepancy in the solution can be explained by Fig. 8b, where the PiNN solution is more diffusive than the FD solution. PiNNs without Fourier Features fail to represent the more heterogeneous problems because the networks return diffusive answers. This shows the strong advantage of PiNNs including Fourier Features especially when the problem is strongly heterogeneous.
Another undesirable behavior present in the PiNNs without Fourier Features is shown in Fig. 9. This graph compares the PiNN loss with and without Fourier Features in Problem 2 and Problem 3. In the Problem 3 case, there is a sharp increase in error, and the LBFGS optimizer stops the training. However, in the Problem 2 case, the error   Another aspect of our modifications to the general PiNN algorithm will be demonstrated in this problem. Figure 10a shows the angular flux result of our PiNNs   with 1000 residual points equally sampled through the entire domain. It can be seen that in the optically thin region (x 2 ½1; 5�), the solution of the neural network is somewhat poor. In order to improve performance, we sample points to ensure a certain point density within each mean free path in the domain. The results of this sampling technique are shown in Fig. 10b, where improvement is evident in the optically thin region. However, some error is still noticeable. To reduce it, we take into account the relative influence caused by an error in the different regions of the problem. For example, an error in the optically thick (x 2 ½0; 1�) region would contribute roughly 50 times more to the loss than the same error in the optically thin region would. To counteract this, we sample more heavily in the optically thin region, so the residuals are both able to be reduced at the same rate, as described in Sec. III.F. The results of the zone-influence sampling are shown in Fig. 10c, where the aggregation of these sampling heuristics is used, leading to good agreement between the PiNNs and the FD solutions.

IV.D. Problem 4
This problem is one of the more difficult to represent with a neural network due to the large void present on the right side of the domain. The void in this problem differs from the void in Problem 5 due to the length of the void region compared to the nonvoid one being much larger in this problem. The extent of each spatial zone and the material properties are given in Table IV. The scalar flux of the PiNN solution and FD solution are shown in Fig. 11. In this problem, the angular direction that starts in the void region should stay at zero for a long length before entering the region with a source on the left side  In order to demonstrate that the length of void is what makes this problem challenging to learn and causes error buildup due to streaming, we present the results of a modified problem with a reduced length of void. The modified problem is described in Table V

IV.E. Problem 5
This heterogeneous problem is a fairly standard benchmark for 1-D transport solvers. A description of the problem is given in Table VI. The results of the PiNNs compared to the results of a FD solution are shown in Fig. 13. There is good agreement between these two solutions. This problem again proves that our PiNN implementation, with Fourier Features, wrapped inside a SI, leads to very good agreement when compared to a fine-mesh FD solution. In Fig. 14, we present the SI   error. This error was calculated by the L-2 relative error between the new and old scalar fluxes.

V. CONCLUSIONS
In this paper, we described and employed PiNNs to solve the radiation transport equation. PiNNs are a type of neural network that uses the residual of a governing equation to train without requiring pregenerated training data. We applied PiNNs to transport problems in heterogeneous 1-D slabs. The standard SI technique was wrapped around the training of the PiNN. We observed that PiNNs can yield poor solutions in heterogeneous problems, and we showed how Fourier Features and improved point-sampling methods can lead to better PiNN performance in those problems. Fourier Features generally improve the PiNN accuracy for heterogeneous problems, and the sampling methods further increased this performance at material interfaces and in problems with very different material properties.
There is a large amount of potential future research into PiNNs applied to radiation transport. First, the integration of scattering into the residual could allow for only a single PiNN to simulate an entire problem with direction as an input. Second, the PiNN algorithm should be tested for radiation problems in higher spatial dimensions. Third, another unique application of this PiNN algorithm is having multiple angular quadratures for a single problem. This is not possible with traditional numerical methods, but PiNNs could be adapted fairly seamlessly such that different spatial regions of the problem have varying levels of angular requirement. Fourth, an investigation into parameterized PiNNs, where material properties become additional inputs to the neural network, would be beneficial. This could reduce the issue of the large computational cost associated with training a PiNN by making the trained PiNN able to very quickly solve a wide range of problems. Fifth, the last potential area of research for PiNNs applied to radiation transport is enforcing causality during training (see, for instance, Ref. 11); this is of interest for radiation transport due to its similarity with the transportsweep solution technique used in traditional S N transport solvers.

Disclosure Statement
No potential conflict of interest was reported by the author(s).