Next Article in Journal
Cognitive Performance and Health-Related Quality of Life in Patients with Neuromyelitis Optica Spectrum Disorder
Next Article in Special Issue
A PDE Model of Breast Tumor Progression in MMTV-PyMT Mice
Previous Article in Journal
Effects of Complex Rehabilitation Program on Reducing Pain and Disability in Patients with Lumbar Disc Protrusion—Is Early Intervention the Best Recommendation?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tracing and Forecasting Metabolic Indices of Cancer Patients Using Patient-Specific Deep Learning Models

1
Beijing Computational Science Research Center, Beijing 100193, China
2
Department of Therapeutic Radiology, Yale University School of Medicine, New Haven, CT 06510, USA
3
Department of Mathematics, University of South Carolina, Columbia, SC 29208, USA
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2022, 12(5), 742; https://doi.org/10.3390/jpm12050742
Submission received: 6 April 2022 / Revised: 25 April 2022 / Accepted: 27 April 2022 / Published: 2 May 2022
(This article belongs to the Special Issue Data-Driven Modeling for Chronic Diseases)

Abstract

:
We develop a patient-specific dynamical system model from the time series data of the cancer patient’s metabolic panel taken during the period of cancer treatment and recovery. The model consists of a pair of stacked long short-term memory (LSTM) recurrent neural networks and a fully connected neural network in each unit. It is intended to be used by physicians to trace back and look forward at the patient’s metabolic indices, to identify potential adverse events, and to make short-term predictions. When the model is used in making short-term predictions, the relative error in every index is less than 10% in the L norm and less than 6.3% in the L 1 norm in the validation process. Once a master model is built, the patient-specific model can be calibrated through transfer learning. As an example, we obtain patient-specific models for four more cancer patients through transfer learning, which all exhibit reduced training time and a comparable level of accuracy. This study demonstrates that this modeling approach is reliable and can deliver clinically acceptable physiological models for tracking and forecasting patients’ metabolic indices.

1. Introduction

With advancements in medical research and practices, more and more cancer patients can live a long time after their cancer treatments, making the quality of life and toxicity management during and post-treatment the primary focus of healthcare providers and cancer patients. How to detect early health anomalies and predict potential adverse effects from some measurable biomarkers or signals for an individual cancer patient is of great importance in cancer patient care [1,2]. The latest Cancer Moonshot 2.0 initiative announced by President Biden aims at reducing the death rate of cancer by 50% in 25 years, which once again put a spotlight on the early detection, diagnosis, and management of cancer [3]. Developing the ability to track a patient’s health status and to monitor the evolution of the specific disease intelligently would provide an enormous benefit to both the patient and the healthcare provider, enabling faster responses to deal with adverse effects and more precise and effective treatments and interventions. With the longitudinally collected time-series data of the patient (e.g., metabolic panel, blood panel test, X-ray and CT scans, etc.) at multiple time points before, during, and after cancer treatments, it is becoming increasingly plausible to have an intelligent tool or device for continuous monitoring, tracking, and forecasting of cancer patients’ health status based on statistical, causal, and mechanistic modeling of patient phenotypes and various biomarkers in the time series data [1,2,4,5,6,7].
The concept of digital twins in healthcare is emerging as a promising platform to develop such a model/tool and gaining a great deal of traction in cancer research and, more broadly, the healthcare community lately [8,9]. The theme in digital twins is to develop a replicate of the targeted health state of the individual patient with an underlying disease based on the patient medical history (medical record, diagnosis, treatments, etc.), the evolutionary trajectory of the disease, and the current health status of the patient. The digital twin model will not only replicate the targeted patient’s state of health relevant to the underlying disease at any given historical time point, but also be able to make a future inference or prediction for the patient’s state of health. Retroceding the patient health state back in time is a distinctive feature that a digital twin has while any live human being does not. With the digital twin’s monitoring, tracking, and predictive capability, physicians can examine the past treatments virtually and explore various treatment pathways/strategies in the future by conducting virtual experiments to come up with an optimal treatment plan/strategy for any specific patient. With this, the digital twin can become a de facto intelligent aide to assist physicians in making decisions in their diagnostic processes and the design of future treatment plans, pathways, or recovery therapies. As a fundamental component in the digital twin model, a physiological module capable of time-reversal monitoring, tracking, and forward prediction of the metabolic panel or the blood panel is indispensable [10].
Recent advances in deep learning applications in healthcare have received a great deal of attention and achieved stunning results in many areas, especially in imaging processing. There have been numerous efforts on deep-learning-based diagnoses using imaging classification [11,12,13]. From an evolutionary perspective, any human body is a dynamical system sustained essentially by food and water intake, as well as air inhalation, while being bombarded incessantly by various external disturbances. Any portion of the human body such as the blood, metabolism, or an individual organ forms its own living subsystem, which perhaps is unavoidably coupled to the entire body’s dynamical system. By ignoring the coupling between the subsystem and the whole body, we assume there exists a self-sustained dynamical system such as the metabolic system in any human body. Any metabolic panel taken from a patient at a given time point would provide a glimpse of the state of the subsystem at that time. With sufficiently collected time series longitudinal data, one would be able to establish a fairly reasonable dynamical system model to describe the underlying dynamics within the subsystem. Such a dynamical system model would possess desired time reversibility in that one can use it to track the previous states in the past and to predict future events as well.
In this study, we aim to develop an AI-enabled deep learning model to monitor, track, and predict a cancer patient’s health status reflected through the metabolic panel from the patient medical records. It is expected to be used by physicians to predict the patient’s treatment outcome trajectory, to reduce the treatment-related adverse effects in the short term, and to improve the quality of life of the cancer patient in the long term. This dynamic model will be patient-specific and updated whenever new health data of the patient become available. A long-short-term memory (LSTM) recurrent neural network (RNN) with a new coupled cell architecture is devised and trained by machine learning based on the time series data in the metabolic panel of cancer patients treated at the Yale-New Haven Health System. We demonstrate how to develop the data-driven, patient-specific model from a set of selected indices from the metabolic panel taken from a cancer patient [14,15,16,17,18,19].
To demonstrate the model’s generalizability to other cancer patients, we recalibrated the model parameters against the metabolic data of the new patient using transfer learning. We believe this modeling approach provides a reliable technical framework for building the physiological module in any digital twin models in healthcare [20].

2. Materials and Methods

We first describe the patient data and data preprocessing used in machine learning and then detail the various LSTM models we developed in this study.

2.1. Data Acquisition and Preprocessing

With the institutional review board’s approval (Yale HIC#1604017609), five cancer patients aged 8 to 17 years old, who have received radiation treatments at Yale-New Haven Hospital between 2013 and 2021, were identified and their metabolic panel results were extracted from the EPIC electronic medical record (EMR) system per the HIPPA regulations. Nineteen metabolic indices were measured during each metabolic procedure for all the cancer patients: Glucose, BUN, Creatinine, BUN/Creatinine Ratio, Anion Gap, CO2, Chloride, Sodium, Potassium, Calcium, Total Protein, Albumin, Globulin, A/G Ratio, Aspartate Aminotransferase (AST), Alanine Aminotransferase (ALT), AST/ALT Ratio, Alkaline Phosphatase, and Total Bilirubin. A series of metabolic panel assessments was performed on each patient at multiple time points (ranging from 20 to 50 times) before, during, and after their cancer radiotherapy.
We selected 9 biomarkers/indices in the patients’ metabolic panel data with the minimal missing data rate as the time series dataset to establish the time-dependent discrete dynamical system. These indices include Glucose, BUN, Creatinine, Anion Gap, CO2, Chloride, Sodium, Potassium, and Calcium. The chosen data for the first patient, whom we name the original patient, was acquired on 42 consecutive days spanning the period from 10 May 2013 to 14 May 2014. We filled the few missing entries in the dataset by using the mean of the two nearest neighbors, i.e., if index x i is missing, we “assign it” the value of x i 1 + x i + 1 2 . The missing data involve the Creatinine index on 26 June 2013 and the Calcium index on 2 January 2014, 3 January 2014, and 4 January 2014, respectively. After making up the missing data points, we computed the mean μ and standard deviation σ of the entire 42 data points for every index. If an index value was larger than μ + 2 σ or less than μ 2 σ , we replaced it by μ + 2 σ or μ 2 σ , respectively, in the dataset to rule out the so-called outlier effect.
Since the clinical data were acquired at non-uniform time intervals, to derive an approximate discrete dynamical system to describe the time series using recurrent neural networks, we needed to generate more data with equal time intervals that included the initial 42 9-dimension/index clinical dataset. We used the first-order linear interpolation to to obtain 739 9-dimensional data vectors, in which adjacent data points are separated from each other by 0.5 days. The correlation coefficients of the 9 indices at the 739 data points are tabulated in Table 1. It shows that no pair of indices in the dataset was highly correlated with a correlation coefficient ≥80%. Hence, we will built the discrete dynamical system using all indices as the input variables of the dynamical system. We chose the LSTM RNN as the framework to build the discrete dynamical system from the preprocessed dataset.

2.2. One-Step Predictive LSTM Model

An LSTM model provides a versatile recurrent neural network architecture with a great deal of flexibility to overcome the gradient vanishing and explosion problems inherent in the conventional recurrent neural networks, while capturing transient dynamics underlying the time series data [19,21,22]. When designing its architecture for our applications, we paid close attention to the input and output data structure to make sure they could describe the underlying time-dependent dynamics since the input and output of an LSTM model do not need to be in the same data format or structure as those in the original metabolic dataset. The general structure of an LSTM cell is shown in Figure 1, where x t is the input to the recurrent neural network cell, c t is the state in the LSTM to enhance the memory effect, and h t is the output of the cell.
A generic LSTM cell is given by the following mathematical formula, where ⨀ indicates the Hadamard product.
f t = σ W f · h t 1 , x t + b f , i t = σ W i · h t 1 , x t + b i , c ˜ t = tanh W c · h t 1 , x t + b c , c t = f t c t 1 + i t * c ˜ t , o t = σ W o h t 1 , x t + b o , h t = o t tanh c t , y t = σ W h t .
In the LSTM cell, f t is called the forget gate and i t the input gate. We multiply the cell state c t 1 by the forget gate f t to control the propagation of the previous information. Then, we add i t c ˜ t to update the current cell state c t . In this process, we combine the previous information with the new information to obtain the final current cell state. Then, we calculate h t and output y t from c t and the previous hidden state h t 1 .
In our design of the model, we used a stacked LSTM architecture coupled with a fully connected neural network for each cell. Figure 2 is a schematic portrait of the stacked LSTM architecture that we adopted in our LSTM model.
The pair of LSTM cells stacked in series and connected to a fully connected output neural network was aimed to achieve a better memory effect. In fact, we can stack more LSTM cells intercalated with fully connected neural networks to form a more complex composite LSTM cell, in which the final output layer taking output h t as the input is a fully connected neural network.
For time series data x 1 , x 2 , , x n , where x i represents the i-th 9-dimensional data vector from the original dataset, and a given time step T > 0 , we concatenate the input 9-dimensional metabolic index data points into a large vector z ( i ) = x i , x i + 1 , , x i + T 1 to define our input to the stacked LSTM cell and define the output of the LSTM cell as y ( i ) = x ^ i + T . The data structure is depicted in Figure 3. The number of total new input–output data pairs ( z ( i ) , y ( i ) ) in the new data structure is N = n T , given as follows
z ( 1 ) = x 1 , x 2 , , x T , y ( 1 ) = x ^ T + 1 , z ( 2 ) = x 2 , x 3 , , x T + 1 , y ( 2 ) = x ^ T + 2 , z ( N ) = x N , x N + 1 , , x T + N 1 , y ( N ) = x ^ T + N .
We designed the LSTM with time step T, input vector z ( i ) and output y ( i ) , i = 1 , , N . By applying the LSTM model through the concatenated dataset ( z ( i ) , y ( i ) ) , i = 1 , , N as an RNN, we predict the next output using the previous T input vectors from the original dataset. We thus name this the one-step predictive LSTM model.
The loss function in the model is defined by
L o s s = 1 M i = 1 M x ^ T + i x T + i 2 2 9 ,
where M is the number of output data vectors in the batch of data.
For the 739 9-dimensional data points in the original time series, we first divided them into the training set and test set in a 9:1 ratio sequentially. With this division, the number of data points in the training set and test set was 665 and 74, respectively. For the data sets, we carried out the zero-mean standardization in each index. Namely, for the j-th index, we computed the mean μ j and unbiased standard deviation σ j of the 665 data points in the training set. Then, we standardized the training data and test data as follows:
x ˜ i j = x i j μ j σ j ,
where i represents the i-th data point and j represents the vector’s j-th entry. Then, we used the standardized training data for the model. The number of input and output pairs of the LSTM model is N = 739 T , and the data pairs are given by ( z ( i ) , y ( i ) ) , i = 1 , , N . The number of training data pairs is 665 T given by ( z ( 1 ) , y ( 1 ) ) to ( z ( 665 T ) , y ( 665 T ) ) . The number of the data pairs used for the test is 74 ranging from ( z ( 666 T ) , y ( 666 T ) ) to ( z ( 739 T ) , y ( 739 T ) ) .
First, we set a loss tolerance ϵ in the training of the neural network. If the training loss was lower than ϵ , we saved the current model parameters and applied the model to the test data. After a prescribed number of epochs, we chose the model that gave the best outcome over the test data as our chosen model. Through an extensive experimentation, we adopted the following hyperparameters for the LSTM model as tabulated in Table 2.
The time step is one of the key hyperparameters in the LSTM model. We trained the model with respect to time step T = 1 , , 10 . The final value of the step is determined by the one that gives the best performance. To evaluate the performance of the model, we define the following metrics:
R M S E = 1 n i = 1 n ( y i y ^ i ) 2 , N S E = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i μ ) 2 , μ = 1 n i = 1 n y i , M A P E = r e l a t i v e L 1 e r r o r = 1 n i = 1 n y i y ^ i y i , r e l a t i v e L e r r o r = max 1 i n y i y ^ i y i .
Note that the LSTM is a versatile recurrent neural network (RNN); there is a great deal of flexibility in its design, especially in its input and output data structures. We showcase some other LSTM designs here and compare their performance with the previous one.

2.3. Input-Stacked LSTM Model

Given the versatility of an LSTM design, we can stack the input and output vectors to form larger input and output vectors. For example, when stack number S = 3 , we first stack data vectors as follows: X i = [ x i , x i + 1 , x i + 2 ] T , where x i is the i-th 9-dimensional data vector. Then, the input and output data pairs are defined in the following forms:
z ( 1 ) = X 1 , X 2 , , X T , y ( 1 ) = [ x ^ 1 + S , x ^ 2 + S , , x ^ T + S ] , z ( 2 ) = X 2 , X 3 , , X T + 1 , y ( 2 ) = [ x ^ 2 + S , x ^ 3 + S , , x ^ T + 1 + S ] , z ( N ) = X N , X N + 1 , , X T + N 1 , y ( N ) = [ x ^ N + S , x ^ N + 1 + S , , x ^ T + N 1 + S ] .
The loss for a batch of output data is defined by:
L o s s = 1 N t = 1 N p = 1 T x ^ p + t 1 + S x p + t 1 + S 2 2 9 · T .

2.4. Multistep Predictive Model

In the models discussed above, prediction step L = 1 . Next, we examine how well the model performs if we extend the one-step predictive model to multiple steps L > 1 by revising the design. Here, the input–output data pairs are defined as follows:
z ( 1 ) = x 1 , x 2 , , x T , y ( 1 ) = x ^ T + 1 , x ^ T + 2 , , x ^ T + L , z ( 2 ) = x 2 , x 3 , , x T + 1 , y ( 2 ) = x ^ T + 2 , x ^ T + 3 , , x ^ T + L + 1 , z ( N ) = x N , x N + 1 , , x T + N 1 , y ( N ) = x ^ T + N , x ^ T + N + 1 , , x ^ T + N + L 1 .
The hyperparameters used are again the ones in Table 2.

3. Results

We present the results obtained using the LSTM models for the original patient alluded to in the previous section firstly. Then, we showcase the results of the patient-specific model for each of the other four patients obtained from transfer learning.

3.1. One-Step Prediction

We considered both the r e l a t i v e L 1 e r r o r and the r e l a t i v e L e r r o r of the nine indices when choosing the best time step T for the LSTM RNN. The errors of the LSTM model with T = 1 , , 10 , when applied to the test set, are listed in Table 3. It follows from Table 3 that the LSTM with time step T = 2 gives the best one-step prediction. We note that the dimensional results are converted from the standardized, dimensionless quantities as follows:
y ^ i j = y ¯ i j · σ j + μ j ,
where y ¯ is the dimensionless quantity used in the neural network model.
The results of the one-step predictive model with T = 2 are depicted in Figure 4, where the maximum relative error is about 9% and the average relative error is about 1.5%. These results are clinically acceptable. Table 4 lists the relative errors of each index for the LSTM model with T = 2 . Except for the largest relative error at about 9%, all others are significantly less than 9%, showing that the one-step predictive model is fairly accurate. When we varied the time step T, we noticed that the LSTM model in fact gave the smallest average error at time step T = 1 ; however, its absolute error was larger than the model with T = 2 . As T increased, the average error increased as well, and so did the absolute error. There was an exception at T = 6 though, where the absolute error was much smaller than the model at T = 5 and T = 7 . As a result, we chose T = 2 as the hyperparameter of the LSTM model in making one-step predictions.

3.2. Input-Stacked LSTM Model

For time step T = 1 , , 5 , respectively, we varied stack number S from one to four to examine the effect of the stack number on the outcome of the LSTM model. The other hyperparameters of the input-stacked LSTM are given in Table 2. We show the minimum training loss after 300 epochs in Table 5 with respect to different stack numbers. The results in Table 5 show that the fitting capacity of the stacked LSTM improved as the number of stacks increased. This is because, when the number of stacks increases, more historical information from the input is added to improve the training loss. For the generalization error of the trained model over the stacked test dataset, we computed the average relative L 1 error of the nine indices for a set of selected stack numbers. The results are shown in Table 5 as well. Notice that the generalization error of the model increased as the stack number increased, indicating increased overfitting in the training of the LSTM model as S increases since the input of the training set and testing set can be quite different after all. This study indicated that using the input-stacked LSTM may not gain any advantages in improving generalization errors over the non-stacked one alluded to earlier, at least for the dataset we used.

3.3. Multistep Prediction

We tabulate the results in the largest L 1 relative error and the largest r e l a t i v e L e r r o r for L = 1 , 2 , 3 , 4 with respect to T = 1 , 2 , 3 , 4 , 5 in Table 6, respectively. From the results in the table, we concluded that time step T = 1 performed the best in this LSTM model, in which the average error was in general not amplified much; however, the relative L norm grew significantly as L increased, especially for large L. The results of the LSTM with T = 1 and different prediction step L = 1 , 3 are depicted in Figure 5 and Figure 6, respectively, as two examples.
In principle, we can use the LSTM model with L > 1 for multistep predictions. However, the predictive power deteriorates rapidly while L increases. For instance, the largest r e l a t i v e L e r r o r was about 10.8 % at L = 3 , and it increased to 14.7% at L = 4 . The average errors in the L 1 norm were much smaller than those in the L norm. The largest L 1 error at L = 4 was less than 4.2% in the study.

3.4. Model Comparison

In one-step predictions, we found that the input-stacked LSTM model improved the model’s fitting capacity at the expense of decreased computational efficiency, although the fitting capacity of the one-step predicative LSTM model was sufficient for the current problem without using stacked input data. However, while fitting capacity improved, the generalization error in the input-stacked LSTM model may deteriorate as the number of the predicative steps increases. Therefore, our assessment of the two LSTM models is that the one-step predicative LSTM model without stacked input data is simpler and perhaps the better model to use for the given dataset.
In multistep predictions, we directly added multiple output steps. After an extensive search for the hyperparameters, we can extend the prediction steps to L = 3 under the restriction that the largest relative L loss among all indices is around 10%. We saw relative errors larger than 10% for steps beyond L = 3 and, therefore, would advise against using the model beyond L = 3 for the given dataset.

3.5. Transfer Learning of LSTM Models to Fit Other Patients

Given the patient-specific LSTM models presented above, we next discuss how to apply the models to other patients through transfer learning. In transfer learning, the hyperparameters of the deep neural network model are kept so that the training of the deep neural network is much more efficient in terms of epochs. The patient-specific models for the four other patients through transfer learning showed similar accuracy in short-term predictions, demonstrating the effectiveness and reliability of this approach. Specifically, we employed transfer learning to retrain the model parameters (weight and biases) to fit the same set of indices from the metabolic panel of the other four cancer patients, coded respectively as ESK, MS, PH, and SCC, while retaining the hyperparameters of the LSTM model.

3.5.1. One-Step Prediction

For the four additional patients, we fixed the hyperparameters of the LSTM model given in Table 2 and then retrained the weight and biases of the LSTM using the parameters in the already trained model as the initial guess. The patient-specific model for each patient was trained in a much-reduced number of epochs. The corresponding relative L 1 and L errors for all four patient-specific LSTM models are tabulated in Table 7. The models for patients ESK and SCC trained especially well with relative errors better than the ones from the model for the original patient. The “worse” and the “worst” case were given by the models for patient PH and MS, respectively. If we examine the model for PH closely, we notice that the L 1 error in the model was consistently less than 8.6% for all indices, while the L error of BUN was about 17.5%, Anion Gap was about 11.22%, and others were all less than 9.4%; the NSE value corresponding to BUN and Anion Gap was 0.612 and 0.045, respectively, indicating the fitting was performed reasonably well, although they were not accurately reflected by the relative errors. Analogously, we noticed that the L 1 error for BUN in the model for patient MS was 32%, while the L was over 100%. This is because the relative error metric used in this case involves a small number in the denominator, where an absolute error metric is perhaps more accurate than the relative error metric. The actual NSE value for BUN in this model was 71.42%. Therefore, the fitting result is not that bad.
The input and output data and the one-step prediction for patients ESK, PH, SM, and SCC are plotted in Figure 7, Figure A1, Figure A2 and Figure A3, respectively. The specific models for patients ESK and SCC made better predictions than the original one. If we examine the absolute error, the model for patient MS was no worse than the one for patient ESK. We depict Figure 7 in the text here and put Figure A1, Figure A2 and Figure A3 in the Appendix A.

3.5.2. Multistep Prediction

Then, we checked the performance of the models in making multistep predictions, where the hyperparameters of the LSTM models are given in Table 2 and the prediction step L = 3 . We plot model outputs for time step L = 3 in Figure 8, Figure 9, Figure A4, and Figure A5, respectively. The corresponding relative errors are tabulated in Table 8. The results are comparable to those in the one-step prediction if we take the relative error as the evaluation criterion: we can achieve good multistep predictions on patients ESK, PH, and SCC, while we cannot predict as well on patient MS. However, the prediction made by the model for patient MS performed reasonably well in BUN, as shown in Figure 9, if other assessment metrics besides the relative errors are taken into account. We show Figure 8 and Figure 9 here and put Figure A4 and Figure A5 in the Appendix A.

4. Discussion

A human being is a complex dynamical system. Hence, the indices in the patient’s metabolic panel follow an intrinsic evolutionary path in time, which is undoubtedly influenced by many health factors within the human body. As a first step toward establishing the digital twin model for the patient, it would be reasonable to assume that the indices in the metabolic panel form a self-consistent dynamical system. Under this assumption, we embarked on a journey to build a discrete dynamical system model to approximate the underlying dynamics of the patient based on an LSTM RNN. The model leverages the historical data acquired in time series to infer the future behavior of the indices in the metabolic panel in time. The rationale for adopting the LSTM architecture to build the discrete dynamical system model is because of its versatility in the structure and functional design to mimic a discrete dynamical system using deep neural networks. We note that we have had some successful experience with continuous dynamical system modeling of time series data in the past for albumin dynamics coupled with several other biomarkers using neural ordinary differential equations [7]. We tried the previous approach for the given datasets. Unfortunately, that approach did not produce a desirable result for the time series data in the metabolic panel that we studied.
In this study, we were limited by the availability of accurate raw data in all indices. Therefore, we only applied the modeling framework to a subset of the indices in the metabolic panel to showcase the effectiveness of the modeling approach and the usefulness of the resulting models. We believe we can readily extend this to all indices in the metabolic panel should a complete dataset be available. In the model developed for the first cancer patient, the one-step prediction amounted to about 12 h or 1/2 day. In both the absolute value metric ( L norm) and the average L 1 norm, the relative error was less than 10%. These error ranges are in general acceptable clinically. When the five models were applied to make one-step predictions, the trend was generally predicted correctly and the errors were in acceptable ranges. The model is expected to give physicians a reasonable short-term prediction with a correct trend and a reasonably close value to assist their decision-making in treatment planning, medical history tracking, and monitoring.
From all the numerical experiments conducted, we noticed that the multistep predictions of the models were in general less accurate than those in the one-step predictions. This is because the model we used relies heavily on adequate and accurate data, but there is inevitably noise in the data collected from the patients over time and the amount of data from each patient is relatively small, limited by the current medical practices and regulations. For the medical data, how to reduce the data noise and improve the robustness of the data-driven modeling is still a challenging problem to be solved. We noticed that there were some existing predictive models for other medical issues in the literature that could obtain a smaller relative L 1 error (e.g., lower than 1%) [23,24]. The temporal fluctuations in the metabolic panel data we acquired were more significant. We should not expect the same level of error in the outcome of the predictive model. In addition, we added the feasibility study of transfer learning of the deep learning model to other patients of the same disease and established the corresponding clinically acceptable results. If we want to apply the patient-specific model in the real-world scenario, the model must have the ability to be transferred to new patients.
Moving forward, we plan to apply the developed LSTM RNN framework to predict the risks of stroke and heart disease for individuals based on their time series metabolic biomarkers. This future work is clinically important, as heart diseases and strokes are the #1 and #5 causes of death in the United States, respectively [25]. That our LSTM model can accurately predict the personal metabolic variations some period ahead is a critically important feature of this model as the early warning and prevention procedures could largely reduce the risk of heart attack and stroke, hence saving lives and improving the quality of life for millions of people worldwide. We will report the results of this new development in our future work.
In addition, we will explore the possibility of developing separate dynamical models based on an individual’s other blood test results, a procedure often performed in the clinic to evaluate one’s overall health status and detect a wide range of disorders such as anemia, infection, and leukemia. Similar to the comprehensive metabolic panel (CMP), there is a series of complete blood count (CBC) results registered for individual patients in the electronic medical record (EMR) system. However, the biomarkers are different with different correlations between them; hence, different machine learning algorithms may be needed to better describe the dynamics of health status using these time series CBC data. Furthermore, we will test if the higher dimensions of patient data by combining CMP and CBC biomarkers could lead to better prediction over a 12 h’ early warning time. All these efforts will eventually lead to the ultimate physiological digital twin module we would like to develop for a patient, not only to replicate the patient physiological state from the past to now, but also to make reasonable inferences for the near future.
Limited by sparsity in the time points when the longitudinal data were collected, continued assessment of the model on a larger cohort of cancer patients and assessment of model robustness on augmented longitudinal data collected at refined time points are necessary. The lack of statistical details in the data collection points in the training dataset may also limit the inference capability of the model. All these are challenging issues that need further refined analyses.

5. Conclusions

We developed a discrete dynamical system model describing the time-dependent dynamics of the metabolic panel of a cancer patient using the LSTM recurrent neural network architecture. The patient-specific model can be used to make short-term predictions in one step with relative errors consistently less than 10% in the absolute value metric and much less than 10% in an average sense. It can be applied to make multistep predictions with a slightly elevated error level (i.e., relative error less than 11% in three steps and 15% in four steps). Using four additional cancer patients’ metabolic panels, we show that the patient-specific LSTM model can be retrained through transfer learning. This modeling platform has great potential for identifying potential dynamical features in the metabolic panel of cancer patients and patients of other prominent diseases, and thereby serves as an important module in the more general digital twin for the cancer patient.

Author Contributions

Conceptualization, Q.W.; Methodology, J.H., C.L. and Q.W.; Software, J.H.; Validation, J.H., J.D. and Q.W.; Data Curation, J.D.; Writing—Original Draft Preparation, J.H. and C.L.; Writing—Review & Editing, J.D. and Q.W.; Supervision, J.D. and Q.W.; Project Administration, Q.W.; Funding Acquisition, J.D. and Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

Jianguo Hou acknowledges the support through NSFC award NSAF-U1930402 to CSRC. Jun Deng, Chunyan Li, and Qi Wang’s research is partially supported by the DOE (DE-SC0021655) and NCI via Leidos Biomedical Research, Inc. (21X130F), awards on Digital Twins. Chunyan Li and Qi Wang also acknowledge partial support from NSF awards DMS-1954532 and OIA-1655740 and a GEAR award from the SC EPSCoR/IDeA Program.

Institutional Review Board Statement

This study was carried out with the approval of Yale institutional review board (HIC#1604017609).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy issues.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Other Plots of the Transfer Learning

The corresponding output plots of the other patients are shown below.
Figure A1. The one-step prediction of the LSTM model for patient MS with time step T = 2 .
Figure A1. The one-step prediction of the LSTM model for patient MS with time step T = 2 .
Jpm 12 00742 g0a1
Figure A2. The one-step prediction of the LSTM model for patient PH with time step T = 2 .
Figure A2. The one-step prediction of the LSTM model for patient PH with time step T = 2 .
Jpm 12 00742 g0a2
Figure A3. The one-step prediction of the LSTM model for patient SCC with time step T = 2 .
Figure A3. The one-step prediction of the LSTM model for patient SCC with time step T = 2 .
Jpm 12 00742 g0a3
Figure A4. The multistep prediction of the LSTM model for patient PH with T = 1 , L = 3 .
Figure A4. The multistep prediction of the LSTM model for patient PH with T = 1 , L = 3 .
Jpm 12 00742 g0a4
Figure A5. The multistep prediction of the model for patient SCC with T = 1 , L = 3 .
Figure A5. The multistep prediction of the model for patient SCC with T = 1 , L = 3 .
Jpm 12 00742 g0a5

Appendix A.2. Statistical Difference

The mean and standard deviation of the raw data from the five patients, including training and testing data, are tabulated in Table A1 and Table A2, respectively. The mean values of the patient data were relatively close, but the variances were different, indicating data fluctuations among different patients.
Table A1. Mean values of the indices of in the patients’ data.
Table A1. Mean values of the indices of in the patients’ data.
IndexInitialESKMSPHSCC
Glucose101.37599.527598.60494.160298.7027
BUN13.4736.4619510.343211.571310.827
Creatinine0.4817320.4775850.4519140.4236410.366111
Anion Gap11.62339.2969811.965611.139311.3348
CO223.905924.898923.539223.835624.8449
Chloride102.826105.838104.09101.897103.923
Sodium138.207140.124139.6136.927140.11
Potassium3.774873.721493.940114.140174.06711
Calcium9.602119.051129.400429.631249.79586
Table A2. Standard deviations in the data of the five patients’.
Table A2. Standard deviations in the data of the five patients’.
IndexInitialESKMSPHSCC
Glucose12.258219.000913.412610.328213.3031
BUN3.623261.427374.136053.00012.40213
Creatinine0.05657060.04681060.09140020.0722390.0360429
Anion Gap1.149821.647352.191551.644261.78242
CO21.806191.574072.390492.02761.16956
Chloride2.343361.758862.932313.073831.23467
Sodium1.234961.350682.023571.987121.26889
Potassium0.2685890.1841930.4804370.3371780.189848
Calcium0.3242990.3009060.348670.3704160.15704

Appendix A.3. Performance Measures of All One-Step LSTM Models

We define the absolute mean error as
M A E = 1 n i = 1 n | y i y ^ i | .
The detailed error estimates of the patient-specific, one-step prediction LSTM models are tabulated in the following five tables with time step T = 2 . From the tables, we find the model works well with the MAPE evaluation criterion. Because of the intrinsic problem in the formula of the NSE, the results of the NSE can be bad when the data are close to a constant. Hence, we focus on the results given by the MAPE.
Table A3. Errors in other metrics of the one-step prediction of the original patient.
Table A3. Errors in other metrics of the one-step prediction of the original patient.
IndexRMSEMAEMAPENSE
Glucose2.5151.9460.02088−0.2631
BUN0.94320.87260.06344−0.9996
Creatinine0.0044530.0037440.0068060.9747
Anion Gap0.17920.16170.015750.9555
CO20.33390.26190.010340.5067
Chloride0.30510.23590.0022670.8091
Sodium0.23740.20380.0014650.9576
Potassium0.056160.049400.012730.02123
Calcium0.023760.020490.0020980.7503
Table A4. Errors in other metrics of the one-step prediction of patient ESK.
Table A4. Errors in other metrics of the one-step prediction of patient ESK.
IndexRMSEMAEMAPENSE
Glucose5.0224.4220.044500.3181
BUN0.29290.27070.046110.9428
Creatinine0.012300.012120.023080.3026
Anion Gap0.19510.14590.015150.5312
CO20.12370.10740.0041820.9798
Chloride0.11710.092190.00087510.8365
Sodium0.080010.056140.00039990.9317
Potassium0.025490.022630.0059590.9867
Calcium0.024290.019160.0021230.5892
Table A5. Errors in other metrics of the one-step prediction of patient MS.
Table A5. Errors in other metrics of the one-step prediction of patient MS.
IndexRMSEMAEMAPENSE
Glucose4.7914.1290.044350.3112
BUN1.8421.2960.32670.7141
Creatinine0.041240.028670.055830.3251
Anion Gap0.98170.60570.044350.6150
CO20.83990.60800.028600.2224
Chloride1.4111.0220.0097410.8205
Sodium1.2070.91440.0065420.6265
Potassium0.26680.21830.065210.5498
Calcium0.19380.13920.015720.6879
Table A6. Errors in other metrics of the one-step prediction of patient PH.
Table A6. Errors in other metrics of the one-step prediction of patient PH.
IndexRMSEMAEMAPENSE
Glucose3.4752.6510.030490.8609
BUN1.4851.2620.086000.6123
Creatinine0.027520.024170.04714−0.2638
Anion Gap0.80560.67340.052600.04500
CO20.56650.46980.021410.7005
Chloride0.80550.68430.0066510.6286
Sodium0.49400.38720.0028170.8139
Potassium0.12000.11330.026230.5025
Calcium0.095820.084120.0088220.5527
Table A7. Errors in other metrics of the one-step prediction of patient SCC.
Table A7. Errors in other metrics of the one-step prediction of patient SCC.
IndexRMSEMAEMAPENSE
Glucose0.31430.30380.0031830.9106
BUN0.13270.12770.0088910.9840
Creatinine0.0071480.0067130.016450.3344
Anion Gap0.035270.030600.0028540.9594
CO20.038990.032660.0012850.9876
Chloride0.029920.021870.00020890.9927
Sodium0.0070900.004883 3.47 × 10 5 0.9983
Potassium0.0022210.0018470.0004617 2.36 × 10 26
Calcium0.0090610.0089600.00092090.7326

References

  1. Henry, N.L.; Hayes, D.F. Cancer biomarkers. Mol. Oncol. 2012, 6, 140–146. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Srinivas, P.R.; Srivastava, S.; Hanash, S.; Wright, G.L., Jr. Proteomics in early detection of cancer. Clin. Chem. 2001, 47, 1901–1911. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Biden, J. Fact Sheet: President Biden Reignites Cancer Moonshot to End Cancer as We Know It. The White House Statements and Release, 2 February 2022. [Google Scholar]
  4. Wu, L.; Qu, X. Cancer biomarker detection: Recent achievements and challenges. Chem. Soc. Rev. 2015, 44, 2964–2997. [Google Scholar] [CrossRef] [PubMed]
  5. Pearl, J. Causal inference in statistics: An overview. Stat. Surv. 2009, 3, 96–146. [Google Scholar] [CrossRef]
  6. Rubin, D.B. Causal inference using potential outcomes: Design, modeling, decisions. J. Am. Stat. Assoc. 2005, 100, 322–331. [Google Scholar] [CrossRef]
  7. Lei, C.; Wang, Y.; Zhao, J.; Li, K.; Jiang, H.; Wang, Q. A Patient-Specific Predictive Model for Human Albumin Based on Deep Neural Networks. Comput. Methods Programs Biomed. 2020, 196, 105555. [Google Scholar] [CrossRef] [PubMed]
  8. Bruynseels, K.; de Sio, F.S.; van den Hoven, J. Digital Twins in Health Care: Ethical Implications of an Emerging Engineering Paradigm. Front. Genet. 2018, 9, 31. [Google Scholar] [CrossRef] [PubMed]
  9. Hempel, V. Healthcare Solution Testing for Future|Digital Twins in Healthcare. Digital Health Network, 29 December 2017. [Google Scholar]
  10. Forsell, K.; Björ, O.; Eriksson, H.; Järvholm, B.; Nilsson, R.; Andersson, E. Cancer incidence in a cohort of Swedish merchant seafarers between 1985 and 2011. Int. Arch. Occup. Environ. Health 2022, 1–9. [Google Scholar] [CrossRef]
  11. Litjens, G.; Sánchez, C.I.; Timofeeva, N.; Hermsen, M.; Nagtegaal, I.; Kovacs, I.; Hulsbergen-Van De Kaa, C.; Bult, P.; Van Ginneken, B.; Van Der Laak, J. Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci. Rep. 2016, 6, 26286. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Vasile, C.M.; Udriștoiu, A.L.; Ghenea, A.E.; Popescu, M.; Gheonea, C.; Niculescu, C.E.; Ungureanu, A.M.; Udriștoiu, Ș.; Drocaş, A.I.; Gruionu, L.G.; et al. Intelligent Diagnosis of Thyroid Ultrasound Imaging Using an Ensemble of Deep Learning Methods. Medicina 2021, 57, 395. [Google Scholar] [CrossRef] [PubMed]
  13. De Fauw, J.; Ledsam, J.R.; Romera-Paredes, B.; Nikolov, S.; Tomasev, N.; Blackwell, S.; Askham, H.; Glorot, X.; O’Donoghue, B.; Visentin, D.; et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018, 24, 1342–1350. [Google Scholar] [CrossRef] [PubMed]
  14. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef] [PubMed]
  15. Gers, F.A.; Eck, D.; Schmidhuber, J. Applying LSTM to time series predictable through time-window approaches. In Neural Nets WIRN Vietri-01; Springer: Berlin/Heidelberg, Germany, 2002; pp. 193–200. [Google Scholar]
  16. Gers, F.A.; Schraudolph, N.N.; Schmidhuber, J. Learning precise timing with LSTM recurrent networks. J. Mach. Learn. Res. 2002, 3, 115–143. [Google Scholar]
  17. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. A comparison of ARIMA and LSTM in forecasting time series. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 1394–1401. [Google Scholar]
  18. Karevan, Z.; Suykens, J.A. Transductive LSTM for time series prediction: An application to weather forecasting. Neural Netw. 2020, 125, 1–9. [Google Scholar] [CrossRef] [PubMed]
  19. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; The MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  20. Hernandez-Boussard, T.; Macklin, P.; Greenspan, E.; Gryshuk, A.; Stahlberg, E.; Syeda-Mahmood, T.; Shmulevich, I. Digital twins for predictive oncology will be a paradigm shift for precision cancer care. Nat. Med. 2021, 27, 2065–2066. [Google Scholar] [CrossRef] [PubMed]
  21. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning with Applications in R; Springer: London, UK, 2017. [Google Scholar]
  22. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  23. Shang, C.; Gao, J.; Liu, H.; Liu, F. Short-Term Load Forecasting Based on PSO-KFCM Daily Load Curve Clustering and CNN-LSTM Model. IEEE Access 2021, 9, 50344–50357. [Google Scholar] [CrossRef]
  24. El-Sappagh, S.; Abuhmed, T.; Islam, S.R.; Kwak, K.S. Multimodal multitask deep learning model for Alzheimer’s disease progression detection based on time series data. Neurocomputing 2020, 412, 197–215. [Google Scholar] [CrossRef]
  25. Kochanek, K.; Xu, J.; Arias, E. Mortality in the United States, 2019. NCHS Data Brief 2020, 395, 1735–1780. [Google Scholar]
Figure 1. Schematics of an LSTM cell.
Figure 1. Schematics of an LSTM cell.
Jpm 12 00742 g001
Figure 2. Schematics of an RNN cell consisting of a pair of stacked LSTM cells and a fully connected neural network at the output layer. The part in the rectangular box bounded by the dashed lines is optional.
Figure 2. Schematics of an RNN cell consisting of a pair of stacked LSTM cells and a fully connected neural network at the output layer. The part in the rectangular box bounded by the dashed lines is optional.
Jpm 12 00742 g002
Figure 3. Schematics of the input and output data format/structure.
Figure 3. Schematics of the input and output data format/structure.
Jpm 12 00742 g003
Figure 4. Output { y ^ i , i = 1 , , } of the LSTM with time step T = 2 . The values used in training are plotted on the left of the vertical line, while the predicted ones are on the right. The largest L 1 error is about 1.5%, and the largest L error is about 9%.
Figure 4. Output { y ^ i , i = 1 , , } of the LSTM with time step T = 2 . The values used in training are plotted on the left of the vertical line, while the predicted ones are on the right. The largest L 1 error is about 1.5%, and the largest L error is about 9%.
Jpm 12 00742 g004
Figure 5. Output { x ^ T + L + i 1 , i = 1 , , } at T = 1 , L = 1 . The ones used in training are plotted on the left of the vertical line and the predicted ones on the right.
Figure 5. Output { x ^ T + L + i 1 , i = 1 , , } at T = 1 , L = 1 . The ones used in training are plotted on the left of the vertical line and the predicted ones on the right.
Jpm 12 00742 g005
Figure 6. Output { x ^ T + L + i 1 , i = 1 , , } at T = 1 , L = 3 . The ones used in training are plotted on the left of the vertical line and the predicted ones on the right.
Figure 6. Output { x ^ T + L + i 1 , i = 1 , , } at T = 1 , L = 3 . The ones used in training are plotted on the left of the vertical line and the predicted ones on the right.
Jpm 12 00742 g006
Figure 7. The outputs of the one-step predictive LSTM model for patient ESK with time step T = 2 . The values on the right of the vertical line are predicted.
Figure 7. The outputs of the one-step predictive LSTM model for patient ESK with time step T = 2 . The values on the right of the vertical line are predicted.
Jpm 12 00742 g007
Figure 8. Multi-step predictions of the LSTM model for patient ESK at T = 1 , L = 3 . The values on the right of the vertical line are predicted.
Figure 8. Multi-step predictions of the LSTM model for patient ESK at T = 1 , L = 3 . The values on the right of the vertical line are predicted.
Jpm 12 00742 g008
Figure 9. Multistep predictions of the LSTM model for patient MS at T = 1 , L = 3 .
Figure 9. Multistep predictions of the LSTM model for patient MS at T = 1 , L = 3 .
Jpm 12 00742 g009
Table 1. Correlation coefficients of the 9 metabolic indices over the interpolated data points.
Table 1. Correlation coefficients of the 9 metabolic indices over the interpolated data points.
GlucoseBUNCreatineAnion GapCO2ChlorideSodiumPotassiumCalcium
Glucose1
BUN−0.5663361
Creatine−0.2837790.5487641
Anion Gap−0.08715970.4354520.2829751
CO2−0.7442250.7579750.3515490.07992331
Chloride0.602712−0.784478−0.382908−0.368508−0.7922481
Sodium−0.07712870.0227023−0.01178060.1382860.09231140.3737841
Potassium0.0807286−0.543226−0.012367−0.422068−0.477360.52685−0.07057651
Calcium−0.7543210.5755920.3019830.1185040.613964−0.5062140.0840671−0.04045921
Table 2. Hyperparameters of the LSTM model.
Table 2. Hyperparameters of the LSTM model.
Train–Test ProportionBatch SizeLearning RateOptimizerError Tolerance ϵ LSTM DirectionTime Steps
initial patient one-step prediction9:1160.001Adam0.01unidirectional1–10
initial patient input-stacked LSTM9:1160.001Adam0.05unidirectional1–5
initial patient multistep prediction9:1160.001Adam1unidirectional1–5
new patients’ one-step prediction9:1160.001Adam0.01unidirectional2
new patients’ multistep prediction9:1160.001Adam1unidirectional2
Table 3. The R e l a t i v e   L 1   E r r o r and R e l a t i v e   L   E r r o r of the LSTM model at different time steps T.
Table 3. The R e l a t i v e   L 1   E r r o r and R e l a t i v e   L   E r r o r of the LSTM model at different time steps T.
Time Step TRelative L 1 ErrorRelative L Error
T = 10.013800.1141
T = 20.015090.08914
T = 30.015090.1052
T = 40.018110.1264
T = 50.020270.1249
T = 60.020720.09490
T = 70.020740.1327
T = 80.021590.1360
T = 90.022450.1395
T = 100.021890.1388
Table 4. Relative errors for all 9 indices of the metabolic panel in the LSTM model with T = 2 .
Table 4. Relative errors for all 9 indices of the metabolic panel in the LSTM model with T = 2 .
IndexRelative L 1 ErrorRelative L Error
Glucose0.020880.06424
BUN0.063440.08914
Creatinine0.0068060.01754
Anion Gap0.015750.03049
CO20.010340.03068
Chloride0.0022670.007395
Sodium0.0014650.003298
Potassium0.012730.02356
Calcium0.0020980.005678
Table 5. The minimum training loss and average r e l a t i v e L 1 e r r o r at different stack numbers S.
Table 5. The minimum training loss and average r e l a t i v e L 1 e r r o r at different stack numbers S.
Minimum Training LossStack 1Stack 2Stack 3Stack 4
T = 10.033690.027920.021010.01742
T = 20.026160.022400.017120.01454
T = 30.022820.019540.015480.01297
T = 40.020920.017700.014370.01178
T = 50.019560.016360.013340.01091
relative L 1 error Stack 1Stack 2Stack 3Stack 4
T = 10.021540.023120.024840.02427
T = 20.019390.023260.020680.02108
T = 30.016520.022110.022250.02312
T = 40.017080.022770.024080.02556
T = 50.016640.023100.025750.02792
Table 6. The relative L 1 error and the relative L error in multistep predictions.
Table 6. The relative L 1 error and the relative L error in multistep predictions.
L 1 ErrorL = 1L = 2L = 3L = 4
T = 10.013800.017900.022880.02905
T = 20.014310.024650.029300.03480
T = 30.013450.028960.037210.04095
T = 40.012510.031180.040200.04209
T = 50.012170.031270.041370.04174
relative L error L = 1L = 2L = 3L = 4
T = 10.11410.097970.10770.1464
T = 20.09760.26650.27500.2685
T = 30.10390.31770.34780.3584
T = 40.10480.38930.35200.3313
T = 50.090470.40200.37370.2987
Table 7. Relative L 1 and L errors for all 9 indices of the patients.
Table 7. Relative L 1 and L errors for all 9 indices of the patients.
IndexESK L 1 MS L 1 PH L 1 SCC L 1 ESK L MS L PH L SCC L
Glucose0.044500.044350.030490.0031830.087780.10670.088910.003952
BUN0.046110.32670.086000.0088910.062702.7770.17500.01438
Creatinine0.023080.055830.047140.016450.030920.22030.093990.02750
Anion Gap0.015150.044350.052600.0028540.050690.15280.11220.0065
CO20.0041820.028600.021410.0012850.010140.098820.051950.002946
Chloride0.00087510.0097410.0066510.00020890.0026770.031720.051950.0007556
Sodium0.00039990.006540.002817 3.474 × 10 5 0.0015990.019260.0089320.0001211
Potassium0.0059590.065210.026230.00046170.015280.17650.040950.001250
Calcium0.0021230.015720.0088220.00092090.0058220.056460.016630.001075
Table 8. Relative L 1 and L errors for all 9 indices of the patients at L = 3 .
Table 8. Relative L 1 and L errors for all 9 indices of the patients at L = 3 .
IndexESK L 1 MS L 1 PH L 1 SCC L 1 ESK L MS L PH L SCC L
Glucose0.076650.074360.069790.0034860.16910.12590.20920.008344
BUN0.098130.62430.12720.027950.13374.7120.22110.03364
Creatinine0.044920.071190.046590.048290.053240.25660.091570.07156
Anion Gap0.012220.070390.057900.040890.041590.25140.13190.06099
CO20.0060380.039370.039600.010400.023520.15110.12800.01529
Chloride0.0023570.019060.010500.00017420.0055950.043450.029130.0002961
Sodium0.0012210.0080140.0058170.00048710.0029880.023700.019900.0009125
Potassium0.021600.080460.024360.00091670.041920.27030.077690.002582
Calcium0.0026620.026360.013590.0022780.0077070.079770.033650.003785
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hou, J.; Deng, J.; Li, C.; Wang, Q. Tracing and Forecasting Metabolic Indices of Cancer Patients Using Patient-Specific Deep Learning Models. J. Pers. Med. 2022, 12, 742. https://doi.org/10.3390/jpm12050742

AMA Style

Hou J, Deng J, Li C, Wang Q. Tracing and Forecasting Metabolic Indices of Cancer Patients Using Patient-Specific Deep Learning Models. Journal of Personalized Medicine. 2022; 12(5):742. https://doi.org/10.3390/jpm12050742

Chicago/Turabian Style

Hou, Jianguo, Jun Deng, Chunyan Li, and Qi Wang. 2022. "Tracing and Forecasting Metabolic Indices of Cancer Patients Using Patient-Specific Deep Learning Models" Journal of Personalized Medicine 12, no. 5: 742. https://doi.org/10.3390/jpm12050742

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop