Efficient enhancement of low-rank tensor completion via thin QR decomposition

Low-rank tensor completion (LRTC), which aims to complete missing entries from tensors with partially observed terms by utilizing the low-rank structure of tensors, has been widely used in various real-world issues. The core tensor nuclear norm minimization (CTNM) method based on Tucker decomposition is one of common LRTC methods. However, the CTNM methods based on Tucker decomposition often have a large computing cost due to the fact that the general factor matrix solving technique involves multiple singular value decompositions (SVDs) in each loop. To address this problem, this article enhances the method and proposes an effective CTNM method based on thin QR decomposition (CTNM-QR) with lower computing complexity. The proposed method extends the CTNM by introducing tensor versions of the auxiliary variables instead of matrices, while using the thin QR decomposition to solve the factor matrix rather than the SVD, which can save the computational complexity and improve the tensor completion accuracy. In addition, the CTNM-QR method's convergence and complexity are analyzed further. Numerous experiments in synthetic data, real color images, and brain MRI data at different missing rates demonstrate that the proposed method not only outperforms in terms of completion accuracy and visualization, but also conducts more efficiently than most state-of-the-art LRTC methods.


Introduction
With the rapid development of human needs and Internet technology, the scale of data acquired by people has expanded.Tensor data, such as third-order color images, fourthorder video sequences, and hyperspectral images, are now ubiquitous.At present, tensors are widely used in machine learning (Bai et al., 2021;Panagakis et al., 2021), computer vision (Kajo et al., 2019), image processing (Miao et al., 2020), etc., and are better at expressing the complex internal structure of higher-order data than vectors and matrices.However, due to data transmission and storage limitations, practical applications often face missing or corrupted observation tensor entries, which directly affect the quality of tensor analysis.Consequently, utilizing the observation part for missing tensor completion is a prominent research topic.
Conventional methods of completion involve matrixization or vectorization of the tensor, which leads to dimensionality catastrophe in addition to destroying the spatial structure of high-order data.The aim of low-rank tensor completion (LRTC), a multidimensional expansion based on low-rank matrix completion (LRMC) (Candes and Recht, 2012;Xu et al., 2021), is to estimate missing terms (Zhou et al., 2018) by utilizing spatial correlation between tensor terms.LRTC is crucial for tensor processing and analysis.Several researches (Kolda and Bader, 2009;Liu et al., 2013;Zhao et al., 2020) have demonstrated that there is a significant quantity of redundant information in natural tensor data, which is typically low-rank or virtually low-rank, such as photos and videos.Figure 1 displays the distribution of singular values for the three channels of the color image.It can be observed that the majority information of the image is stored in only a few of the bigger singular values, its low rank structure is notable.This tensor property encourages the development of low-rank tensor estimation and completion problems, and it has been proactively applied in real-world problems like multi-channel image and video completion (Liu et al., 2013;Su et al., 2022), audio source separation (Yoshii et al., 2018), etc.
In the two-dimensional case, the rank is a strong tool for capturing global information about the data, but the definition of the tensor rank is not as explicit as that of the matrix, and the fundamental issue is that the definition of the tensor rank is not unique.The CANDECOMP/PARAFAC (CP) tensor rank is defined as the minimum number of rank-1 tensors needed in the CP decomposition.Nevertheless, CP rank determination is an NPhard problem (Hillar and Lim, 2013), and there isn't always an optimal low CP rank approximation, which may result in a poor fit in real-world applications (Qiu et al., 2022).Another is the Tucker rank derived from the Tucker decomposition, defined as a vector composed of the ranks of mode unfolding matrices, which is extensively employed in LRTC driven by its superior computeability.In recent years, tensor train (TT) rank (Oseledets, 2011), tubal rank (Zhang and Aeron, 2016;Zhou et al., 2018), tensor ring (TR) rank (Yuan et al., 2019a), tensor tree rank (Liu et al., 2019b), etc. have also been successively proposed to provide an effective way to deal with the LRTC problem.
Owing to the nonconvex nature of the rank function, contemporary LRTC approaches predominantly employ the tensor nuclear norm as a convex substitute for the rank function.Liu et al. (2013) pioneered the definition of the tensor nuclear norm, modeled the LRTC as a convex optimization problem, and suggested three efficient algorithms to solve the LRTC problem.Shang et al. (2017) introduced a novel method for fuzzy dual nuclear norm minimization.In a related development, Bengua et al. (2017) leveraging the TT rank, proposed the SiLRTC-TT and TMac-TT models.These algorithms have significantly contributed to addressing the LRTC problem.Nevertheless, the majority of algorithms relying on tensor nuclear norm minimization (TNNM) encounter challenges associated with the elevated computational expense incurred by multiple SVDs.In addressing this issue, Liu et al. (2014a) formulates a novel tensor completion approach by imposing nuclear norm constraints on the factor matrices of CP decomposition.Shi et al. (2017) presented a tensor rank estimation approach based on l 1 -regularized orthogonal CP decomposition.Liu (2013) builds upon the TNNM, introduces the Tucker decomposition, and proposes a Core Tensor Nuclear Norm Minimization (CTNM) completion model.
While the CTNM model reduces the computational burden to a certain degree, the introduction of the Tucker decomposition inevitably involves the iterative solution of the factor matrix.A higher-order expansion of the matrix SVD to the tensor, the Higher Order SVD (HOSVD) (De Lathauwer et al., 2000a;Chao et al., 2021) was first proposed.De Lathauwer et al. (2000b) introduced the Higher-Order Orthogonal Iteration (HOOI) approach, which initializes the HOSVD solution, to boost approximation accuracy.These factor matrix solution techniques also involve multiple SVDs of tentor matricizations in the loop iterations, which surely places some computing strain on the CTNM model and contributes to its still-relatively-slow convergence rate.Building upon the insights from Shi et al. (2015), we adopt the thin QR decomposition, as opposed to HOOI, to address the factor matrix.We propose an efficient CTNM-QR model that circumvents the necessity for SVD in factor matrix resolution, thereby enhancing the algorithm's computational productivity.The principal contributions of this article are as follows: (1) An enhanced CTNM-QR model is proposed, via introducing tensor versions of the auxiliary variables instead of matrices, while using the thin QR decomposition to solve the factor matrix instead of HOOI, which avoids the computation of multiple SVDs in each loop and minimizes the computational cost to a certain extent, thus speeding up the process.
(2) The complexity and convergence performance of the proposed method are analyzed.Compared with the state-of-the-art LRTC methods, the proposed method saves computational complexity, and is consistent with the generalized higher-order orthogonal iteration (gHOI) algorithm (Liu et al., 2014b) in terms of the algorithm's primary computational cost.
(3) The CTNM-QR model is applied to synthetic tensors, color images, and medical MRI scans for completion, validating the accuracy and efficacy of the enhanced algorithm.
The rest of the paper is organized as follows.We give the tensor notation and review some related work in Section 2. In Section 3, the proposed CTNM-QR method is described in detail, and the convergence and complexity of the CTNM-QR method are analyzed.In Section 4, we offer experimental analysis and completion results to validate the proposed algorithm.Finally, the paper is summarized in Section 5.

Notations and background
In this section, we introduce some notation and background of the tensor and tensor completion.Throughout this article, we use lowercase letters to denote vectors, e.g., a, b, uppercase letters for matrices, e.g., A, B, and uppercase fancy letters for tensors of order three and above, e.g., A, X.The ith element of a vector a is denoted as a i , the element of a matrix A with index (i, j) is denoted as a ij , and the elements of a N-order tensor

. Tensor concepts and terminology
In this subsection, we first introduce some fundamental tensor concepts and terminology.An extensive review of tensors are discussed in Kolda and Bader (2009).
For the definition of a tensor, there are several definitions at different abstraction levels (Boumal, 2023).Here, an Nth-order tensor is defined as a multidimensional array with N dimensions (N > 2), which is an extension of a matrix to higher order.Here, N represents the order of the tensor and the number of dimensions of the tensor, also known as modes.Analogous to the rows and columns of a matrix, we can create subarrays by fixing some of the given tensor's indices.A fiber of a tensor is a vector created by fixing all but one index of a particular mode (Filipović and Jukić, 2015).A third-order tensor X has row, column and tube fibers, denoted by x(:, j, k), x(i, :, k), x(i, j, :), respectively.A slice of a tensor is defined as a two-dimensional matrix, obtained by fixing all but the indices of two particular modes, e.g., a third-order tensor with horizontal, lateral, and frontal slices, denoted by x(i, :, :), x(:, j, :), and x(:, :, k), respectively.
We can also transform a tensor into a matrix via reordering the elements of an Nth-order tensor, known as matricization, unfolding or flattening.The mode-n unfolding of an Nth-order tensor X ∈ R I 1 ×I 2 ×•••×I N is obtained by keeping the nth mode fixed and concatenating the slices of the rest of the modes into one long matrix, denoted by X (n) , which is an I n × j =n I j matrix for n = 1, 2, • • • , N, where the symbol is continued product.The tensor element (i 1 , i 2 , . . ., i N ) is mapped into the element i n , j of the matrix X (n) , where If there exists a series of distinct modes, the order of multiplication in the mode-n product is independent, that is, According to the mode-n unfolding of the tensor, we have the following formula The inner product of two tensors with the same size is denoted as a sum of the products of the corresponding elements, that is, Frontiers in Big Data frontiersin.orgWu and Jin ./fdata. .
By an extension of the matrix property, the Frobenius norm of the tensor X ∈ R I 1 ×I 2 ×•••×I N is defined as the square root of the sum of the squares of all elements: The distance between the tensor X and Y is defined as X − Y F .The Kronecker product of two matrices A ∈ R I×J and B ∈ R K×L is denoted by A ⊗ B, and yields a matrix of size (IK) × (JL), defined as: .

Tensor decomposition
Tensor decomposition refers to expressing a tensor as a sequence of elementary operations on other simple arrays.There are many tensor decompositions.We introduce two major tensor decompositions, including CANDECOMP/PARAFAC (CP) decomposition (Cattell, 1944) and Tucker decomposition (Tucker, 1963).
For a given tensor A ∈ R I 1 ×I 2 ×•••×I N , the CP decomposition is defined as the factorization of A into the minimum number of linear combinations of rank-1 tensors: where "•" denotes the outer product of vectors, and The Tucker decomposition is based on an extension of the CP decomposition, a form of higher-order principal component analysis.The Tucker decomposition of the tensor A is expressed as: is the core tensor and denotes the degree of interaction between the different components, and U n ∈ R I n ×R n is the nth-mode factor matrix, denoting the principal components on each mode, usually orthogonal.In general, R n is much smaller than I n for n = 1, • • • , N, so the core tensor C can be considered as a compressed version of the original tensor A, which reduces the computational complexity to some extent.
The matricized form of the Tucker decomposition is as follows: The Tucker-rank of the Nth-order tensor A ∈ R I 1 ×I 2 ×•••×I N , also known as the multilinear rank (Kasai and Mishra, 2016), is denoted as an N-dimensional vector composed of the ranks of the various mode unfolding matrices of the tensor: where rank A (n) is the rank of the mode-n unfolding matrices, representing the dimension of the vector space spanned by the mode-n fiber.In contrast to the CP rank, the Tucker-rank can be easily obtained by computation, while the CP rank can only be determined empirically or by multiple experimental comparisons to search for the optimal value (Mu et al., 2014).Diagram of the Tucker decomposition of a third-order tensor is shown in Figure 2. .

Low-rank tensor completion
The LRTC aims to infer and fill missing entries of the tensor from partially observed values via utilizing the low-rank structure of the tensor, which is an extension of the low-rank matrix completion.
Let M ∈ R p×q be the observed incomplete matrix and be the index set, i.e., the observed positions of the matrix elements.
Recovering the matrix M based on the observed partial elements is referred to as low-rank matrix completion, that is, where the complete matrix X ∈ R p×q is obtained via the above optimization problem with constraints.
As a broadening of the low-rank matrix completion (Equation 2), the LRTC (Ji et al., 2009;Liu et al., 2013;Yuan et al., 2019a) is formulated as the following optimization problem: is the incomplete tensor containing the missing elements, and X is the completed tensor.The optimization problem is built with a rank minimization constraint, which ensures that the output tensor has the lowest rank possible for tensor completion.In the LRTC, there is also a category of optimization based on tensor decomposition, formulated as minimizing the discrepancy between the real observations and the predicted values to complete the missing entries under the given tensor rank.That is, where R is a bounded constraint on the rank of the X and the tensor X fulfills a specific tensor decomposition model.The traditional models are the completion models based on the CP and Tucker decompositions, e.g., the optimization model for tensor completion using CP decomposition is stated as follows: The LRTC algorithm minimizing tensor rank is more widely used because the low-rank decomposition-based completion algorithm is impractical in some situations in reality, especially it is difficult to select an appropriate tensor rank for use in the algorithm when there are few observed elements.However, since the tensor rank is not unique, this makes the LRTC algorithm very challenging (Liu et al., 2013).Besides the CP rank and the Tucker rank in Equation ( 1), there are various forms of tensor rank, such as TT rank, tubal rank, TR rank, etc. Selecting one of these forms will result in different LRTC models.

. Tensor nuclear norm minimization
In the LRTC, since the computation of the tensor CP rank is an NP-hard problem (Hillar and Lim, 2013) and the Tucker rank has the excellent property of being easy to compute, optimal completion models based on the Tucker rank of the tensor are widely used (Romera-Paredes and Pontil, 2013;Liu et al., 2019a;Yu et al., 2023).That is, min where the parameters α i are pre-specified weights and satisfy α i ≥ 0, α i = 1.The Tucker rank of the tensor is represented in the form of a weighted sum.Gandy et al. (2011) suggests the unweighted model of Since the rank of the matrix is non-convex, it is difficult to guarantee to finding a globally optimal solution in the optimization (3).Since the nuclear norm of the matrix is shown to be the tightest convex envelop of the matrix rank (Candes and Recht, 2012), the completion model based on the nuclear norm minimization is proposed as follows: where the nuclear norm X (i) * = j σ j (X (i) ), and σ j (X (i) ) denotes the j-th largest singular value of the matricization X (i) .The N i=1 α i X (i) * is defined as the nuclear norm of the tensor X, denoted as X * .The tensor nuclear norm is also known as the tensor schatten-q norm (Ji et al., 2009;Signoretto et al., 2010).Based on the nuclear norm, the concept of lowrank matrix completion is generalized to high-dimensional data, where the low-rank property of the tensor is often used as a necessary assumption to limit the degrees of freedom of missing entries (Kressner et al., 2014).
The tensor nuclear norm minimization completion model (TNNM) (4) can be expressed as the following: where λ > 0 is a regularization parameter.
Since the matrix nuclear norm in the model ( 5) is interdependent and cannot be optimized independently, auxiliary variables for splitting need to be introduced for simplifying solution algorithm.Liu et al. (2013) presented three solution algorithms, which are simple LRTC, fast LRTC, and high-accuracy LRTC.Based on tensor decomposition, Liu (2013); Liu et al. (2014a) proposed completion models for the factor matrix nuclear norm minimization and the core tensor nuclear norm minimization.Nevertheless, the majority of TNNM algorithms involve SVDs of several large matrices in each iteration, resulting in high computing costs.

The proposed method . Core tensor nuclear norm minimization based on thin QR decomposition
In the subsection, we describe the proposed CTNM-QR algorithm and the solution procedure in detail.Specifically, we adopt the ideas of the thin QR decomposition to update the factor matrix on the framework of the core tensor nuclear norm minimization (CTNM) and propose an efficient low-rank tensor completion algorithm.
In the CTNM model based on the Tucker decomposition (Liu, 2013;Liu et al., 2014b), the following assumption is stated: and the orthogonal factor matrix U i ∈ R I i ×R i .Then where X * and C * , denotes the nuclear norm of the original tensor and the core tensor, respectively.
According to the Assumption 1, the completion model for the CTNM is: where I I i is an I i × I i Identity matrix.Owing to the interdependence of the unfolding matrices, we introduce auxiliary tensors Then, the model ( 6) is transformed into the following equivalent form: For the optimization problem Equation ( 7), we propose an efficient CTNM-QR algorithm based on alternating direction multiplication method (ADMM) to address it.The ADMM is a commonly employed optimization method that decomposes a problem into a sequence of sub-problems to calculate the optimal solution, and is adept at proficiently addressing optimization problems featuring multiple non-smooth terms in the objective function (Glowinski, 2014;Boţ and Nguyen, 2020;Han, 2022).The partial augmented Lagrangian function for the model ( 7) is as follows: where is the Lagrange multiplier tensor and µ > 0 is the penalty parameter.
The ADMM converts the original tensor completion problem (7) into solving the five subproblems of (9a-9e) by minimizing the augmented Lagrangian function ( 8), which iteratively updates the corresponding parameters.Specifically, we update sequentially in the order of (9a-9e) with the fixed other parameters.The steps of the algorithm solution are as follows: That is, for V k+1 i in the k + 1th iteration, (9a) can be written as the following optimization subproblem: The iteration for V k+1 i for i = 1, 2, • • • , N can be effectively addressed through the application of the singular value thresholding (SVT) operator (Cai et al., 2010).It is easy to obtain a closed solution to the problem (10) as: where fold i denotes the transformation of the matrix into a tensor along the ith mode, if the SVD of a matrix For C k+1 in the k + 1th iteration, the optimization problem (9b) with respect to C can be rewritten as: With other variables fixed, it is easy to see that Equation ( 12) is a smooth and differentiable optimization problem that can be derived by first-order optimality conditions for C (Liu, 2013;Liu et al., 2014b).
For X k+1 in the k + 1th iteration, solving X by the following subproblem: By deriving the Karush-Kuhn-Tucker (KKT) condition for Equation (13), it is easy to obtain the update form of X as X k+1 = T and , where is the index set of the observed elements, c is the complement of .
For the factor matrix U k+1 i ∈ R I i ×R i (i = 1, 2, . . ., N) in the k+1th iteration, the optimization (9c) is equivalent to the following optimization problem: This is a peculiar Tucker decomposition model in which U i , i = 1, 2, . . ., N is an orthogonal matrix.HOSVD (De Lathauwer et al., 2000a;Chao et al., 2021) and HOOI (De Lathauwer et al., 2000b) are frequently used in the computation of Equation ( 14).Nevertheless, these approaches need alternating iterations of solving Equation ( 14) and involve SVDs of multiple tensor matricizations in each loop iteration, which causes the algorithms to converge quite slowly.To enhance the approaches, we adopt thin QR decomposition instead of HOOI for solving U i , which essentially avoid the SVDs during factorization and improve the computing efficiency.
Specifically, we first isolate the ith factor matrix U i from the Tucker decomposition since the order of multiplications in the tensor mode-i product is not relevant.Then the equivalent form of the optimization model ( 14) for the solution of the ith factor matrix U i is as follows: Frontiers in Big Data frontiersin.orgWu and Jin ./fdata. .
where the given tensor The problem ( 15) is nonconvex for all variables, yet it is convex for each block tensor.When the variables C, X, {U j } N j=1,j =i are fixed, the optimization (15) with respect to the factor matrix U n is a smooth and differentiable optimization problem and can be solved via the first-order optimality condition.Hence the following equation holds: and Z k i(i) are the mode-i unfolding matrices of the tensors Z k i × i U i − X k and Z k i , respectively.Then the least-squares solution without taking into account the orthogonality constraints on the factor matrices can be obtained by Motivated by the work of Shi et al. (2015), we adopt the thin QR decomposition for the orthogonality constraints of the factor matrices, that is, where QR(•) denotes the thin QR decomposition of the matrix.At this step, the proposed CTNM-QR algorithm merely needs to carry out QR decomposition of a matrix with the While the computational complexity of the HOSVD algorithm in terms , where R i ≪ I i .This indicates that the significant computational complexity associated with the SVD can be effectively mitigated by employing QR decomposition to resolve the factor matrix.

. Implementation
In the following, we discuss several implementations issues.To effectively solve the objective function ( 8), the penalty parameter µ is adaptively varied via setting the initialization of µ to µ 0 , and iteratively increasing µ k by µ k+1 = ρµ k , where ρ ∈ (1.0, 1.1] and µ 0 is usually a smaller constant.For the iteration among the problems (9a-9e), we principally use two convergence criteria.One is the relative change satisfaction between two neighboring recovery tensors X k+1 and X k : where k is the number of iterations and tol is the tolerance error.The alternative halting condition is k ≥ K, where K is the maximum number of iterations.The iteration ends if any of the two conditions is met.We summarize the implementation of the CTNM-QR as Algorithm 1. Update X k+1 by solving Equation ( 13); Update µ k+1 by µ k+1 = min(ρµ k , µ max ); 15 Check the convergence conditions
. Convergence analysis X k be a sequence generated by the Algorithm 1.Although the CTNM-QR is a nonconvex ADMM algorithm, the subsequence are convergent and converge to the KKT point of the problem (6) proved by Liu et al. (2014b).In the subsection, we only illustrate the convergence of the improved part of the algorithm for the subproblem (9c).
Denote the objective function in Equation ( 15) as f . Then the following equation holds:

. Complexity analysis
In the subsection, we analyze the complexity of the proposed CTNM-QR algorithm.The running time of the CTNM-QR algorithm is primarily consumed in solving the auxiliary variables with the SVT operator and some multiplicative computations.Suppose the size of the input tensor is , and the time complexity of computing U n in Equation ( 16) is O NR N i=1 I i + R 2 i I i .Consequently, the total time complexity of the Algorithm 1 is O T(N + 1)R N i=1 I i , where T is the total number of iterations.When the N and I i , i = 1, 2, • • • , N are fixed, the time complexity of the CTNM-QR algorithm mainly dominates on R. Fortunately, the R generally takes small values due to the low-rank assumption of the tensor.
For comparison, we list the major computational complexity of several classical tensor-completion algorithms in Table 1.Although the per-iteration complexity of TT-SGD (Yuan et al., 2019b) is incredibly low, the total number of iterations required is higher and more time-consuming than the proposed CTNM-QR algorithm in practice.In addition, the Polak-Ribiere nonlinear conjugate gradient algorithm of CP-WOPT (Acar et al., 2011) also suffers from the problem of being time-consuming.Compared with the nuclear norm minimization algorithms such as SiLRTC, HaLRTC, CTNM, etc., the proposed CTNM-QR algorithm saves some computational complexity.The CTNM-QR is consistent with the generalized higher-order orthogonal iteration (gHOI) algorithm (Liu et al., 2014b) in terms of the algorithm's primary computational cost.

Experiments
In this section, we conduct experiments to evaluate the performance and efficiency of the CTNM-QR algorithm in the LRTC problem using synthetic tensor data, color images, and medical MRI scans.The experiment compares the CTNM-QR with five other typical completion algorithms: gHOI (Liu et al., 2014b), CTNM (Liu, 2013), FaLRTC (Liu et al., 2013), CP-WOPT (Acar et al., 2011), and TT-SGD (Yuan et al., 2019b), where the parameters of the comparison algorithms are optimized.In addition, by means of studies, we analyze the convergence of the CTNM-QR algorithm intuitively.All the experiments are implemented on Windows 10 and MATLAB (R2018b) with an Intel(R) Core(TM) i5-7200U CPU at 2.70 GHz and 4 GB RAM.
To construct the incomplete tensor artificially, we utilize random uniform sampling from the complete tensor based on the given missing rate (MR), so the tensor missing form is randomized uniform missing.We set up varying MR for the test tensor, where the MR is defined as: where M is the number of entries observed in the N order tensor We adopt relative squared error (RSE) and CPU time as evaluation metrics to measure the performance of the tensor completion.The RSE is calculated as follows: where • F indicates the Frobenius norm, the X and T denote the tensor recovered by the algorithm and the original true tensor, respectively.The algorithm's accuracy in recovering the tensor increases as the RSE decreases, and computational complexity decreases as the CPU time decreases.

. Synthetic tensor completion
In the simulation studies, we generate the low rank synthetic tensor and their entries independently follow the standard normal distribution.This implies that the Tucker rank of the tensor T is (R 1 , R 2 , . . ., R N ).
In the concrete implementation, we set the tolerance to tol = 10 −5 for all algorithms, the maximum number of iterations to K = 6 × 10 5 for the TT-SGD, and K = 500 for the rest of the algorithms.For the FaLRTC, gHOI, CTNM, and CTNM-QR, we set the weights The smoothing parameter of the FaLRTC is set to α i /I i , i = 1, . . ., N, and the rest of the parameters are maintained at their default values.For the CP-WOPT, the tensor rank R is set to 15.For the TT-SGD, the TT rank is set to For the gHOI, CTNM, and CTNM-QR, we set µ 0 = 10 −3 , µ max = 10 10 , ρ = 1.05.
We consider two cases including the fourth-order tensor with the size 40 × 40 × 40 × 40 and the fifth-order tensor with the size 20 × 20 × 20 × 20 × 20.The Tucker-rank of the tensor is set to R 1 = • • • = R N = 5, where the tensor orders N are 4 and 5, respectively.We consider four distinct MRs 30, 50, 70, 80% for the test tensor and replicate 100 simulations.independent experiments.The outcomes reveal that the CTNM-QR method yields more accurate solutions with the less time costs, and outperforms other algorithms in the term of the RSE regardless of whether it's a higher order tensor or a lower order tensor.Although the gHOT method has a somewhat shorter running time than the CTNM-QR at the low MRs, the time costs of both methods are quite close and the RSE of the CTNM-QR is less than one of the gHOI.We also investigate the convergence of the CTNM, gHOI, and CTNM-QR methods on synthetic tensor data with the Tucker-rank (R 1 , R 2 , R 3 ) = (10, 10, 10) and the size 100 × 100 × 100, as shown in Figure 3A, where the abscissa indicates the number of iterations and the vertical indicates the log-value of the relative change of the X k .The results display that the relative change of the convergence of the CTNM-QR drops more quickly and smoothly than that of CTNM and gHOI.Figure 3B illustrates the relative changes of the convergence of the CTNM-QR approach at 40, 60, and 80% MRs.From the Figure 3B, we find that even at the high MR, the CTNM-QR converges fast in fewer than 100 iterations.

. Color image completion
In this subsection, we confirm the efficacy of the CTNM-QR by performing recovery on RGB real-color images, where each image is represented as a third-order tensor.The low-rank nature of color images has been demonstrated in Figure 1, which allows excellent recovery of incomplete images utilizing the low-rank structure of the tensor.The low-rank color pictures chosen for this experiment include Flower, Starfish, Sky, Lena, and Apple.The original images are portrayed in Figure 4.The images are 500 × 500 × 3 tensors and encompass portraits and various natural landscapes.The image pixels are normalized to [0, 1].We carry out multiple experimental comparisons for every image in order to determine the optimal rank value.
In the concrete implementation, the tolerance error of all algorithms is set to tol = 10 −4 , the maximum number of iterations K = 5×10 5 for TT-SGD and K = 500 for the rest of the algorithms.For the FaLRTC, gHOI, CTNM and CTNM-QR, we set the weights α i = [1, 1, 10 −3 ], and the smoothing parameter for the former to α i / √ I i for i = 1, 2, 3.For the CTNM-QR, we set µ 0 = 10 −3 , µ max = 10 10 , ρ = 1.05.The rest of the algorithm parameters are preserved at the default settings.
For the color images, we utilize random uniform sampling to make them missing, and the average reconstruction results from 100 independent experiments, including RSE and time cost (seconds), are reported in Table 3, where MR is set to 40, 60, and 80%.The best results are highlighted in bold in the table, and it is clear that the proposed CTNM-QR method achieves the best RSE values in all images and outperforms the other competing methods in terms of CPU time in a substantial number of cases.For example, when recovering Sky with an MR of 80%, the RSE of the CTNM-QR is improved by 0.036 and 0.071 over gHOI and the original algorithm CTNM, respectively, and by 0.119 over the FaLRTC.In terms of running speed, it outperforms the CTNM and the TT-SGD by around 4 and 7 times, respectively.Furthermore, in the majority of image completions, the CP-WOPT method takes the longest.It is obvious that the CTNM-QR can efficiently mine the low-rank image spatial correlation to precisely recover the missing tensor in a reduced iteration time.
Figure 5 depicts visual comparisons of the observed images with the images following the various completion algorithms, where the missing rate of incomplete images is 60 and 80%, and the missing entries are displayed in black.From Figure 5, it can be seen that the best visibility outcomes are obtained by the CTNM-QR method (Figure 5G), and the recovered image is very clear and close to the original image.In most instances, the results of the CP-WOPT contain some fairly ambiguous parts, and this is the most acute, especially for the Lena and Apple.In the completion of missing data for images with an 80% MR, the FaLRTC CTNM exhibit suboptimal performance, particularly on the Flower and Lena datasets.Although the results of the gHOI and TT-SGD appear to be more discernible in comparison, the finer details of the images have not been fully recovered, indicating a certain degree of information loss.It implies that the proposed CTNM-QR method can recover the images accurately and efficiently.

. Brain MRI completion
In the subsection, we evaluate the performance of the CTNM-QR algorithm via the MRI images completion of the brain.MRI images of the brain are acquired through the use of magnetic resonance imaging, which utilizes a strong magnetic field and harmless radio waves.It has an influential role in the fields of medicine and neuroscience, being a key tool used by physicians to diagnose diseases, study brain activity, and guide treatment programs.Here, we analyze brain MRI data with the size 236 × 236 × 180.Similarly, medical MRI data is approximated as a lowrank tensor and thus can be recovered with low-rank completion algorithms.For comparison, the tolerance value for all algorithms ./fdata. .  is set to tol = 10 −3 , the tensor Tucker-rank is set to R 1 = R 2 = R 3 = 30 for the gHOI, CTNM, and CTNM-QR, the tensor CP rank is set to 40 for the CP-WOPT, and the TT rank is set to R 1 = R 2 = 20 for the TT-SGD.We choose a slice of the MRI images for completion due to their enormous computation cost.Table 4 presents the recovery accuracy (RSE) and time cost (seconds) of every algorithm at varying MRs, where the MRs are set to 30, 50, and 70%, respectively.The findings show that the RSE of the CTNM-QR, gHOI, and CTNM are nearly equal at MR values of 30 and 50%.However, when the MR increases to 70%, the RSE of the CTNM and gHOI dramatically rose.In terms of the runtime, the proposed CTNM-QR method typically runs faster than others, especially more noticeable at the higher MR.In addition, Figure 6 displays the recovery results of the brain MRI slice data at the 70% MR, where the CTNM-QR outperforms the other algorithms in terms of both recovered vision and efficiency.

Conclusion
In this paper, we propose a CTNM-QR method, which is based on an enhancement of CTNM and aims at addressing the problem of completing the low-rank missing tensor.Firstly, we review the CTNM algorithm briefly and discuss our changes, which can be categorized into two major points.One difference is that the model's auxiliary variables are introduced as tensors rather than matrix versions, and the tensor form is able to utilize multidimensional information to better retain the spatial structure of high-dimensional data.The other is that the solution of the factor matrix in the model flexibly adopts the thin QR decomposition instead of HOOI, which avoids the computation of multiple SVDs in each loop and minimizes the computational cost to a certain extent, thus speeding up the process.Secondly, we further investigate the convergence and complexity of the CTNM-QR method.Ultimately, extensive experiments on synthetic data of various orders, real color images, and brain MRI data indicate that our method not only beats most state-of-the-art LRTC algorithms in terms of completion accuracy and visualization, but also runs faster.
In the future, we will continue to explore how to automatically determine the optimal rank of a tensor in CTNM-QR to settle the rank selection challenge.In addition, combining sophisticated techniques such as deep learning will boost the algorithm's ability to adapt to more complicated data.To widen the method's application scenarios, we are interested in adapting CTNM-QR to additional practical applications such as recommender systems, video denoising, and so on.

FIGURE
FIGURE Illustration of the low-rank property of natural color image.(A) A three-dimensional low rank color image X with dimensions × × (taken from https://www.pexels.com/jacquemaydominique).(B) Singular value distribution of the first channel X(:, :, ).(C) Singular value distribution of the second channel X(:, :, ).(D) Singular value distribution of the third channel X(:, :, ).

FIGURE
FIGUREIllustration of the Tucker decomposition of a third-order tensor.
=n is fixed, the objective function f (U n ) tends to be decreasing during the iteration process, and the sequence U k 1 , . . ., U k N is the convergence sequence.

FIGURE
FIGURE Convergence of the algorithm on synthetic data with the size × × .(A) Log-value of relative change of the CTNM, gHOI, and CTNM-QR with % MR. (B) Relative changes of the CTNM-QR with various MRs.

FIGURE
FIGURE Illustration of image completion performance comparisons.The first two rows display the results for a missing rate MR = %, and the last four rows display the results for a missing rate MR = % (best viewed zoomed in).(A) Observation.(B) CP-WOPT.(C) FaLRTC.(D) TT-SGD.(E) gHOI.(F) CTNM.(G) CTNM-QR.Adapted from https://pic.sogou.com.
and the I i × R i Table 2 presents the average results (RSE and time cost) for 100TABLE Average RSE and time cost (seconds) of all approaches on various tensor sizes.Bold values indicate the best results obtained from the experiments.
TABLE RSE and time cost (seconds) comparison on true color images at di erent MRs.Bold values indicate the best results obtained from the experiments.
TABLE RSE and time cost (seconds) comparison on brain MRI scanning slice.Bold values indicate the best results obtained from the experiments.