Next Article in Journal
The First Evidence of the Insecticidal Potential of Plant Powders from Invasive Alien Plants against Rice Weevil under Laboratory Conditions
Previous Article in Journal
High-Quality, InN-Based, Saturable Absorbers for Ultrafast Laser Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Incremental Canonical Correlation Analysis

1
Science and Technology on Parallel and Distributed Processing, National University of Defense Technology, Changsha 410073, China
2
College of Computer, National University of Defense Technology, Changsha 410073, China
3
College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(21), 7827; https://doi.org/10.3390/app10217827
Submission received: 24 September 2020 / Revised: 30 October 2020 / Accepted: 2 November 2020 / Published: 4 November 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:

Featured Application

This paper presents the solutions for real-time visual tasks such as visual tracking. Some real-life scenarios such as tracking specific targets can apply this method by extending ICCA.

Abstract

Canonical correlation analysis (CCA) is a kind of a simple yet effective multiview feature learning technique. In general, it learns separate subspaces for two views by maximizing their correlations. However, there still exist two restrictions to limit its applicability for large-scale datasets, such as videos: (1) sufficiently large memory requirements and (2) high-computation complexity for matrix inverse. To address these issues, we propose an incremental canonical correlation analysis (ICCA), which maintains in an adaptive manner a constant memory storage for both the mean and covariance matrices. More importantly, to avoid matrix inverse, we save overhead time by using sequential singular value decomposition (SVD), which is still efficient in case when the number of samples is sufficiently few. Driven by visual tracking, which tracks a specific target in a video sequence, we readily apply the proposed ICCA for this task through some essential modifications to evaluate its efficacy. Extensive experiments on several video sequences show the superiority of ICCA when compared to several classical trackers.

1. Introduction

Canonical correlation analysis, as a mathematical statistical tool, was proposed by Hovelling [1] in 1936. It is mainly used to analyze the association between two groups of random variables. In 2003, Hardoon et al. [2] briefly reviewed CCA and provided a generalized framework with optimization and kernel tricks. CCA is relevant to mutual information, thus it was used early in information retrieval. Later on, it was broadly applied to reduce dimension [3,4,5], clustering [6,7], regression [8,9], word embedding [10,11,12], and discriminant learning [13,14,15]. In such studies, plain CCA, which was purely used there, ignored the nonlinearity or geometrical structure within data sets. To this end, many variants of CCA emerged in diverse manners. For instance, sparse CCA [16] accounted for feature sparseness for more intuitive interpretation via L 1 regularization. Graph-embedded-based CCA (GCCA) [17,18,19] integrated local manifold structure within data for image classification by graph-embedded subspace learning with auxiliary methods [20,21]. More typically, deep CCA (DCCA) [22] trained the neural network under the guidance of the objective of CCA to model the data nonlinearity. In addition, generalized CCA (GCCA) [23] extended CCA with latent common feature representation. Similar to DCCA, deep GCCA [24] also realized CCA with deep architectures using the objective GCCA. In the above methods, eigenvalue decomposition, matrix inversion, or both are, as usual, involved. This obstructed the applicability of CCA and its variants in large-scale datasets or online learning tasks.
To make CCA efficient, many methods endeavored to explore efficient optimization algorithms, which address the computational inefficiency for the high-dimensional issue of CCA in different settings [25,26,27,28,29,30,31,32,33]. Kettenring et al. [34] showed that CCA was equivalent to the constrained least squares optimization problem. Zha et al. [35] calculated CCA by QR decomposition which denoted that a matrix A ( A R m × n ) was divided into a product of an orthogonal matrix Q and an upper triangular matrix R . However, the large scale of the data would still slow the calculation process. Afterwards, Avron et al. [36] proposed a fast CCA algorithm by employing randomized dimensionality reduction transform [37]. Recently, other methods to solve CCA were based on the methods of gradient descent. Ma et al. [33] introduced the enhanced approximate gradient mechanism and further extended it to stochastic optimization problems. In [38,39], CCA was reformulated into a series of least-square problems, which were solved by fast gradient descent. Such algorithms are still slow in reality. Thus, they are unsuited for online learning tasks.
To address this issue, we develop a simple yet efficient incremental canonical correlation analysis (ICCA) to accelerate plain CCA. Particularly, ICCA maintains constant memory storage for both the mean and covariance matrices, etc., in an adaptive manner. This does not require storage of large amounts of history data, thus a certain space complexity remains. To reduce time complexity, we readily avoid matrix inverse in the learning process by using sequential Karhunen–Loeve (SKL) or singular value decomposition (SVD), which is still efficient in the case that the number of samples is sufficiently few. Driven by visual tracking as an online learning task, which locates a given target in a long video sequence, we apply the proposed ICCA for this task through some essential modifications to ensure its efficacy. Extensive experiments on several video sequences show the superiority of ICCA as compared to several classical trackers. Importantly, the efficiency of ICCA is promising.
There are three aspects to our contribution: (1) we provide a new perspective to solve the efficiency problem of CCA; (2) from the perspective of SVD, we propose an ICCA method to avoid the invert matrix multiplication for saving overhead time; and (3) a tracking algorithm based on ICCA further tries to expand the applicability of CCA for real-life scenarios.
The rest of this paper is outlined as follows. We list the related work about CCA in Section 2. Section 3 details CCA and our method. Section 4 offers the experimental results, and we give our conclusions in Section 5.

2. Related Work

Owing to the problem of CCA described above, a large number of methods were proposed to solve this problem. Most studies were along the specific application of this main line to improve CCA. Based on this consideration, we mainly review in this section those optimization algorithms about CCA that are most related to our method.
To accelerate the efficiency of CCA, a randomized CCA algorithm for a pair of tall and thin matrices was proposed in [36], which first obtained the matrices by randomly reducing the dimensionality of the matrices. The newly generated matrices, as a different dataset, were applied in existing algorithms. As a result, Ma et al. [33] pinpointed that the randomized CCA algorithm still cannot thoroughly solve the original high complexity problems. Lu et al. [31] obtained some suboptimal results by viewing this problem as a sequence of iterative least-squares with imprecise approximation. An iterative algorithm achieved excellent results with a low cost. In [30], an alternating least-squares algorithm was proposed and the performance is better than [33]. Gao et al. [29] found that the objective of CCA is not stochastic convex programming. Ma et al. [33] gave much attention to globally convergent stochastic optimization of CCA. Ge et al. [39] made a breakthrough to this issue. However, the problem of sample complexity still exists. Xu et al. [40] proposed truly alternating least-squares (TALS) for efficient CCA with momentum. The offline algorithm was not suitable for real-time scenes that resulted in high memory storage, so some scholars tried to design fast algorithms to solve the problem of high complexity. Marinov et al. [29] proposed a first-order stochastic approximation algorithm for canonical correlation analysis (CCA), in which the convex relaxation was introduced. Different from these methods, which all approximated the objective of plain CCA, the proposed ICCA directly accelerates original CCA. Note that the sequential Karhunen–Loeve or SVD used in ICCA was originally proposed by Kim [41], which accelerated SVD rather than matrix inversion. Moreover, early on, it was applied in visual tracking [42], which puts forward learning the eigenbasis of PCA. It is obvious that PCA is not CCA and their motivations are very different. More importantly, the proposed ICCA greatly expands the applicability of sequential SVD in multiview tasks.

3. Methodology

3.1. Review of Canonical Correlation Analysis

Canonical correlation analysis (CCA) as a classical multivariate statistical method is widely used to analyze multiview data. In the machine learning community, CCA usually serves as a feature learning method of extracting two separate subspaces to maximize the correlations across views. Consider the samples X R d 1 × n and Y R d 2 × n from two views, where d 1 and d 2 are respectively the dimensions of X and Y , and n denotes the number of the samples; CCA [43] is to seek two subspaces a and b to maximize their projections, as given below:
max a , b a T Σ x y b a T Σ x x a b T Σ y y b
where Σ x y = ( X X ¯ ) ( Y Y ¯ ) T , Σ x x = ( X X ¯ ) ( X X ¯ ) T , Σ y y = ( Y Y ¯ ) ( Y Y ¯ ) T .
To ensure the scale invariance of a and b in Equation (1), the objective function can be rewritten:
max a , b a T Σ x y b s . t .   a T Σ x x a = 1 b T Σ y y b = 1
By constructing a Lagrange equation, Equation (2) is simplified into:
[ Σ x x 1 0 0 Σ y y 1 ] [ 0 Σ x y Σ y x 0 ] [ a b ] = λ [ a b ]
It is proved that the Equation (3) has a global optimal solution, in which a and b are called as canonical subspaces. Before computing a and b , Σ x x 1 , Σ y y 1 , Σ x y , and Σ y x should be calculated at first. However, the time complexity of the matrix inversion about Σ x x 1 and Σ y y 1 is O ( d 3 ) ( d is the larger of the two d 1 and d 2 ), which leads to slow running speed and large calculation overhead [2,40,41]. In practice, there are many data with high dimensions that burden massive data storage. Obviously, using CCA incremental learning is a promising choice to handle this case.

3.2. Incremental Canonical Correlation Analysis (ICCA)

Mean and Covariance Update. In CCA, calculating the mean and covariance is a necessary step for its solution. For the mean and covariance update, it is relatively easy to keep the storage constant. For completeness, we introduce the concrete strategy, where the implementation details could not be slightly different from other methods. Suppose that there are n old samples and m new incoming samples. Let X ¯ and Y ¯ denote the mean of n old samples from X and Y , respectively. Correspondingly, the covariance of n old samples and m new incoming samples between X and Y sample are denoted by Σ x y and Σ x y , respectively. Σ x y denotes the covariance of n t samples.
Σ x y = m 1 n t Σ x y + m 1 n t 1 Σ x y + m ( m + n t ) ( n t 1 ) n t ( X ¯ X ) ( Y ¯ Y ) T
where
Σ x y = i = 1 n ( X i X ¯ ) ( Y i Y ¯ ) T , Σ x y = i = n + 1 n + m ( X i X ¯ ) ( Y i Y ¯ ) T , X ¯ = i = 1 n X i , Y ¯ = i = 1 n Y i , X ¯ = i = n + 1 n + m X i , Y ¯ = i = n + 1 n + m Y i , n t = f n + m
and f denotes a forgotten factor to eliminate old samples, f [ 0 , 1 ] . The parameter f = 1 expresses that more emphasis is put on all the samples, while f = 0 indicates that a full focus is placed on the new samples.
The n previous history matrices including Σ x y , X ¯ and Y ¯ need to be saved before m new samples arrive. The matrix Σ x y , X ¯ and Y ¯ can be obtained with the new m samples and then the covariance of matrix Σ x y can be calculated, based on these n old samples and m new incoming samples. Thus, all the samples do not take up much memory. Of them, the mean of all the samples at this time can be updated as follows:
X ¯ n + m = n n t · X ¯ + n n t X ¯ Y ¯ n + m = n n t · Y ¯ + n n t Y ¯
when the next m samples emerge again, once the matrix Σ x y , X ¯ , and Y ¯ are computed according to Equation (5), the saved value of n t and the historical matrix Σ x y directly take part in the new round of calculation and storage. The complete operation would not increase the storage space with the rapid of the number of samples.
Avoiding Matrix Inversion. To solve CCA, it is necessary to solve the inversion of matrix Σ x x and Σ y y in Equation (3). To avoid the matrix inversion, we resort to the singular value decomposition (SVD), which efficiently solves the thinner matrix. Given any matrix A R d × n , its SVD form is as follows:
A = S V D U S V T
where U R d × d , S R d × n , V R n × n ; then A A T = U S V T V S U T . Owing to the orthogonality V V T = I , and thus A A T = U S S U T . When X X ¯ = U 1 S 1 V 1 T , Σ x x is a real symmetric matrix and thus
Σ x x 1 = ( Σ x x T ) 1 = ( U 1 S 1 2 U 1 T ) 1 = U 1 S 1 2 U 1 T
In Equation (8), there is an inverse operation S 1 2 . Since S 1 is a diagonal matrix whose elements are only located on the diagonal, its inversion can be calculated by the reciprocal of each element. Likewise, Σ y y 1 = ( Σ y y T ) 1 = ( U 2 S 2 2 U 2 T ) 1 = U 2 S 2 2 U 2 T . S 2 2 can be obtained the same as S 1 2 . So, the matrix inverse is avoided.
It is obvious that the overhead of vanilla SVD is not ignored as well, if the matrix is still relatively large. Towards this end, a sequential Karhunen–Loeve algorithm (SKL) [44] can be employed here. Without loss of generality, suppose that n centralized samples X and the m new centered incoming samples X . Assume that X = U 1 S 1 V 1 T . Obviously, the SVD of the concatenation X and X is [ X X ] = U 1 Σ 1 V 1 T . Let X be the component of X orthogonal to U 1 , U = [ U 1 , X ] , and V = [ V 1 0 0 I m ] , where I m is the identity matrix with the size of m . Then, we derive a small matrix X n = U [ X , X ] V T . The SVD of X n is X n = U ˜ D ˜ V ˜ T , then [ X , X ] = U U ˜ D ˜ V ˜ T V T . Obviously, we need the latest left singular matrix U 1 = U U ˜ and the corresponding singular value matrix Σ 1 = D ˜ . Now we can obtain the inverse of the covariance matrix of X .
Adopting the SKL algorithm [44] can ensure its space and time complexity to be constant in O ( n + m ) . Each update only adopts the first top k truncated singular value and basis vectors in the former step and the m new incoming samples, thus the space complexity is reduced to O ( d ( k + m ) ) instead of the previous O ( d ( n + m ) 2 ) as in Levy and Lindenbaum [44].
Updating Subspaces. Of course, the SKL algorithm is designed for sequential SVD methods. Thus, we follow the section above to find the SVD of Σ x y . Since merging the small-sized matrices into the SVD forms of Σ x x 1 , Σ y y 1 , and Σ x y , we can efficiently obtain the SVD of P = Σ x x 1 / 2 Σ x y Σ y y 1 / 2 . In fact, we do not need the concrete form. This is because the projection matrix only requires the left and right singular matrices. The update formula is
{ a = Σ x x 1 / 2 u b = Σ y y 1 / 2 v
where u and v are the left and right singular matrices of P , respectively. Here, the subspace a and b can be updated.
The specific optimization algorithm is shown in Algorithm 1.

4. Experiments

In this section, we apply ICCA for visual tracking with particle filter to illustrate the applicability of incremental CCA (ICCA).

4.1. Settings

Implementation Details. For completeness, following the experimental settings [41], we divide an image into left and right patches. Since occlusion and illumination affects the tracking results, we at the meantime extend the method of [45] to the case of two views. In detail, we directly extract the projection matrices of two views as the bases, which serve as the templates. To integrate two views of templates into the particle filtering framework, we utilize the reconstruction errors to determine the best candidate as the position inference. Different from [45], the updating reconstructed coefficients rely on the templates from two views rather than independent views. For clarity, we give the concrete details as Equation (10):
Algorithm 1. Incremental Subspace Algorithm
Input: n samples X and Y , m new samples X and Y .
Output: the updated subspace N e w _ U 1 , N e w _ U 2 the updated mean and covariance X ¯ n + m , Y ¯ n + m , Σ ˜ x y ;
1. Update the mean X ¯ , Y ¯ and covariance Σ x y and Σ y x of n samples, according to Equation (5);
2. Update the mean X ¯ n + m , Y ¯ n + m of n + m current samples according to Equation (6);
3. X ¯ X ¯ n + m , Y ¯ Y ¯ n + m , Σ ˜ x y Σ x y according to Equations (4) and (5);
4. Calculate U 1 , S 1 , U 2 , and S 2 according to Equation (7);
5. Calculate Σ x x 1 , Σ y y 1 according to Equation (8);
6. Calculate a and b to yield new subspace N e w _ U 1 , N e w _ U 2 according to Equation (9).
min α , e 1 , e 2 1 2 x D 1 α e 1 F 2 + 1 2 y D 2 α e 2 F 2 + μ 1 e 1 1 + μ 2 e 2 2 + λ α 0
where x and y denote the left and right patches, respectively, D 1 and D 2 denote the projection matrices learned by ICCA, α is the sparse coefficient vector, e 1 and e 2 denote the noises, and λ , μ 1 , and μ 2 denote the regularization parameters.
According to [45], we ignore the same solution steps such as the errors e 1 and e 2 , and only describe the reconstructed coefficient update rule. Since the iterative shrinkage technique requires the partial derivative about the coefficient, the first step is to calculate the partial derivative of Equation (10) with respect to α . Let G a is the objective function about α after simplification:
G α = 1 2 D 1 T ( x e 1 ) D 1 T D 1 α F 2 + 1 2 D 2 T ( y e 2 ) D 2 T D 2 α F 2
Then the gradient about α is:
G α α = ( D 1 T D 1 + D 2 T D 2 ) 1 ( D 1 T ( x e 1 ) D 2 T ( y e 2 ) )
Similarly, the gradient about e 1 and e 2 is e 1 ( x D 1 α ) and e 2 ( y D 2 α ) . With the three gradients, α , e 1 , and e 2 can be obtained by combination with APG (accelerated proximal gradient).
In visual tracking with a particle filter, one particle with maximum probability must be chosen as the target location in the next-coming frame. According to Bayesian theory, the maximum probability of particles is expressed as follows:
p ( X i , Y i | S m ) = exp ( J ( α * , e 1 * , e 2 * ) )
where J ( α , e 1 , e 2 ) = 1 2 x D 1 α e 1 F 2 + 1 2 y D 2 α e 2 F 2 + μ 1 e 1 1 + μ 2 e 2 1 , α * , e 1 * , e 2 * are the optimal solution of Equation (10). S m represents the particle state in the m -th frame. These states are composed of six affine parameters and are estimated with the maximum posteriori probability. The maximum probability particle is calculated and treated as the target location in the next frame.
In order to ensure the fairness of comparison and the reliability of the experiment, the running environment of all video frames is performed on the same workstation. In order to fully mine multimode features, the ICCA tracker uses two parts of each particle evenly divided horizontally or vertically as two view features. In the experiment, the regularization parameters of ICCA tracker are set: λ is selected from {0.005, 0.05, 0.06, 0.25, 0.1}, μ 1 is selected from {0.06, 0.08, 0.09, 0.1, 0.15, 0.25, 0.35, 0.3}, and μ 2 is selected from {0.06, 0.08, 0.09, 0.1, 0.15, 0.25, 0.35, 0.3}.
Datasets. This section compares ICCA with several representative trackers, such as robust fragments-based tracking (Frag) [46], visual tracking decomposition (VTD) [47], incremental learning for robust visual tracking (IVT) [42], real time robust L1 tracker using accelerated proximal gradient approach (L1APG) [48], visual tracking with online multiple instance learning (MIL) [49], correlation-based incremental visual tracking (TLD) [41] on a series of video sequences, including basketball, car4, cardark, carscale, deer, Dudek, faceocc1, fish, football, and 18 other video sequences to verify the effectiveness of ICCA. Of the trackers compared, VTD selects multiple image feature models and motion models at the same time, and tracks them respectively. Finally, the optimal feature and motion model are combined as the optimal target area in real time; IVT is a video tracking algorithm based on the incremental update appearance model and maintains a template in the whole tracking process, and dynamically updates the template through incremental PCA, so that it can effectively adapt to changes in appearance. L1APG is based on sparse representation and uses accelerated proximal gradient (APG) to solve the L1 minimization problem. The MIL tracker treats the visual tracking problem as the multi-instance learning problem, in which the bag is composed of many instances instead of a single instance. TLD combines the traditional tracking algorithm and detection algorithm to solve the problems of deformation and occlusion in tracking. Frag is recognized by partial matching. Its target template is described by multiple fragments and blocks of the image. Blocks are arbitrary, not based on the target model. These trackers used here represent different types of trackers and are the most representative shallow learning trackers, which are widely used for comparison purposes in research reports. More importantly, such main tracking algorithms are integrated into a tracking framework [50], which provides the evaluation and figure codes for fair comparison. This paper also implements the ICCA algorithm on the basis of this tracking framework. The video sequences used here, which include different appearance changes of tracking objects in different scenes, can be downloaded from this tracking framework as well.

4.2. Results

In single object tracking, two evaluation criteria are mainly used: qualitative analysis and quantitative analysis. In a video sequence, the tracking effect belongs to the category of qualitative analysis. To track the target, whether the tracking is lost from the first frame to the last frame, and whether the target tracking frame can dynamically follow the change of target size is qualitative analysis. In addition, precision and success rate are used for quantitative analysis. The traditional method to evaluate the tracker is to take the real position of the tracking target in the first frame as the initial tracking target in the test set, and calculate the average accuracy or success rate through the whole test sequence. This is a one-way evaluation (OPE). In addition to these evaluation criteria, we also report the efficiency of each algorithm in terms of the number of frames per second (FPS).

4.2.1. Efficiency Comparison

In Table 1, we offer the FPS values of the compared trackers on all eighteen videos, which are computed as the ratio of the total number of frames in each video length to the total running time, namely, the number of frames per second based on various tracking methods. The larger the value, the higher the efficiency of the tracking algorithm. As in Table 1, ICCA achieves larger values of FPS on most videos. In some videos, like football, jumping, and mhyang, the changes in appearances make the tracking difficult, which leads to the slower convergence of the tracking algorithm based on our ICCA with the same criterion used on the other videos. Thus, the values of FPS are not always the best.

4.2.2. Qualitative Comparison

In the basketball and football videos, the target is a fast-moving person. For the basketball video, when more than one player grabs the ball, the tracking player is easily affected by the background information. In the football video, in addition to being easily affected by the background information, the helmet information is similar, when the players occlude each other, and the tracked target is easy to lose.
Figure 1a,j show good tracking ability of ICCA, which is better than other trackers in the football video. In the three videos of car4, cardark, and carscale, fast-moving cars drive on the streets during the day and night, as well as on the outdoor roads during the day. For the video car4, when the car passes through the overpass or under the shade of trees, its surface light changes greatly. In the video cardark, cars on the street at night are easily blurred by the brightness of other lights or street lights. The car in the video carscale will drive from far to near, and the scale will change when passing through a cluster of branches. Figure 1b–d shows that ICCA tracking results are attractive. In Figure 1e, namely, the video david3, the tracked person walks outdoors. In the process of walking back and forth, it is occluded twice by the trunk. Most tracking algorithms lose the target, while ICCA still works stably. In the deer video sequences, most algorithms fail because the tracked deer is very similar to the surrounding background and moves rapidly; ICCA effectively performs the tracking. The dog in the video Dog1 moves from near to far but slowly, thus most methods can track the target stably. As in Figure 1, the videos faceocc1, faceocc2, and jogging-2 share the common things: the tracked targets are all sometimes occluded, especially the target in jogging-2 has almost complete occlusion, and the color of the occlusion appearance is very similar to the target. ICCA works well and shows its effectiveness. In the video fish, because the camera is shaking slightly, it is more difficult to track the target. ICCA can track the target accurately. In Figure 2a, since the building in the video crossing blocks the sun, the pedestrian tracking is relatively stable in the building shadow. When walking to the shadow and direct light, illumination variations are quite different. In this case, most trackers miss the target, but ICCA can still track the target even when the tracking bounding box size changes a little. In the singer1 video, the background clutter affects the effective learning of the subspace. Likewise, ICCA can track the target correctly.

4.2.3. Quantitative Comparison

To quantify the visual tracking performance of ICCA, seven trackers were evaluated and compared by success rate and precision. It can be seen from Figure 3 that ICCA has almost the best comprehensive performance, and has a high success rate on different attributes, such as background speckle illumination, scale change, rapid movement, and small rotation of objects, but it achieves slightly worse performance than other trackers on out-of-plane rotation, occlusion, and deformation.
In particular, ICCA merely depends on the subspace learning, which as usual is not very robust. ICCA mainly accelerates CCA in an incremental way, in which the solution to CCA is a closed-form rather than stochastic approximation. In addition, ICCA primarily focuses on the efficiency and applicability in real scenarios.
Figure 4 shows the precision comparison of seven trackers over 18 video sequences. ICCA is consistently superior to other trackers for high accuracy on most attributes. Since ICCA just accelerates CCA without improving the robustness of target appearance, the trackers based on ICCA are still sensitive to the occlusion, scale size, and deformation. Thus, on such attributes, the performance based on ICCA is slightly lower than other trackers. We cannot guarantee our tracker to work on all the videos because we only improve the efficiency of CCA rather than its model capacity.
In short, according to the comparison of success rate and accuracy, the ICCA tracker is better than other previous trackers and exhibits better performance on many video sequences.

5. Conclusions

This paper details a simple yet efficient incremental CCA (ICCA), and applies it for visual tracking to verify the effectiveness and promising potential for proposed online learning schemes in reality. Different from existing works, the proposed ICCA directly accelerates the process of CCA, in contrast to some other approximations. The algorithmic efficiency is greatly improved. However, it still needs to develop other applications for further in-depth analysis.

Author Contributions

Conceptualization, H.Z.; methodology, H.Z.; software, H.Z.; validation, H.Z., D.S. and Z.L.; formal analysis, H.Z. and Z.L.; investigation, D.S.; resources, H.Z.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, H.Z.; visualization, H.Z.; supervision, D.S. and Z.L.; project administration, Z.L.; funding acquisition, H.Z. and Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Educational Commission of Hu Nan Province, grant number 16B276.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Harold, H. Relations between two Sets of Variables. Biometrika 1936, 28, 3–4. [Google Scholar]
  2. Hardoon, D.; Szedmak, S.; Shawe-Taylor, J. Canonical Correlation Analysis: An Overview with Application to Learning Methods. Neural Comput. 2004, 16, 2639–2664. [Google Scholar] [CrossRef] [Green Version]
  3. Sun, L.; Ceran, B.; Ye, J.P. A scalable two-stage approach for a class of dimensionality reduction techniques. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 25–28 July 2010. [Google Scholar]
  4. Yuan, Y.H.; Sun, Q.S.; Ge, H.W. Fractional-order embedding canonical correlation analysis and its applications to multi-view dimensionality reduction and recognition. Pattern Recognit. 2014, 47, 1411–1424. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Zhang, J.; Pan, Z.; Zhang, D. Multi-view dimensionality reduction via canonical random correlation analysis. Front. Comput. Sci. 2016, 10, 856–869. [Google Scholar] [CrossRef]
  6. Blaschko, M.B.; Lampert, C.H. Correlational spectral clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 July 2008; pp. 1–8. [Google Scholar]
  7. Chaudhuri, K.; Kakade, S.M.; Livescu, K.; Sridharan, K. Multi-view clustering via canonical correlation analysis. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009. [Google Scholar]
  8. Kakade, S.M.; Foster, D.P. Multi-view Regression Via Canonical Correlation Analysis. In International Conference on Computational Learning Theory; Bshouty, N.H., Gentile, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  9. Lambert, Z.V.; Wildt, A.R.; Durand, R.M. Redundancy analysis: An alternative to canonical correlation and multivariate multiple regression in exploring interset associations. Psychol. Bull. 1988, 104, 282–289. [Google Scholar] [CrossRef]
  10. Dhillon, P.; Rodu, J.; Foster, D.; Ungar, L. Two Step CCA: A new spectral method for estimating vector models of words. Comput. Sci. 2012, 2, 67–74. [Google Scholar]
  11. Gong, Y.; Ke, Q.; Isard, M.; Lazebnik, S. A Multi-View Embedding Space for Modeling Internet Images, Tags, and Their Semantics. Int. J. Comput. Vis. 2014, 106, 210–233. [Google Scholar] [CrossRef] [Green Version]
  12. Nam, K.M.; Song, H.J.; Kim, J.D. Find Alternative Biomarker via Word Embedding. In Proceedings of the Green and Smart Technology, Lisbon, Portugal, 19–22 December 2015; pp. 789–792. [Google Scholar]
  13. Kim, T.K.; Kittler, J.; Cipolla, R. Discriminative Learning and Recognition of Image Set Classes Using Canonical Correlations. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1005–1018. [Google Scholar] [CrossRef] [PubMed]
  14. Su, Y.; Fu, Y.; Gao, X.; Tian, Q. Discriminant Learning Through Multiple Principal Angles for Visual Recognition. IEEE Trans. Image Process. 2012, 21, 1381–1390. [Google Scholar] [CrossRef]
  15. Yi, Z.; Schneider, J.G. Multi-Label Output Codes using Canonical Correlation Analysis. Mach. Learn. Res. 2011, 15, 873–882. [Google Scholar]
  16. Wang, M.; Shao, W.; Hao, X.; Shen, L.; Zhang, D. Identify Consistent Cross-Modality Imaging Genetic Patterns via Discriminant Sparse Canonical Correlation Analysis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 99, 1. [Google Scholar] [CrossRef] [PubMed]
  17. Hu, R.; Zhu, X.; Cheng, D.; He, W.; Yan, Y.; Song, J.; Zhang, S. Graph self-representation method for unsupervised feature selection. Neurocomputing 2017, 220, 130–137. [Google Scholar] [CrossRef]
  18. Li, S.; Zeng, C.; Fu, Y.; Liu, S. Optimizing multi-graph learning based salient object detection. Signal Process. Image Commun. 2017, 55, 93–105. [Google Scholar] [CrossRef] [Green Version]
  19. Peng, Y.; Wang, S.; Long, X.; Lu, B.L. Discriminative graph regularized extreme learning machine and its application to face recognition. Neurocomputing 2015, 149, 340–353. [Google Scholar] [CrossRef]
  20. Li, S.; Tang, C.; Liu, X.; Liu, Y.; Chen, J. Dual graph regularized compact feature representation for unsupervised feature selection. Neurocomputing 2019, 331, 77–96. [Google Scholar] [CrossRef]
  21. Tang, C.; Zhu, X.; Liu, X.; Li, M.; Wang, P.; Zhang, C.; Wang, L. Learning a Joint Affinity Graph for Multiview Subspace Clustering. IEEE Trans. Multimed. 2019, 21, 1724–1736. [Google Scholar] [CrossRef]
  22. Andrew, G.; Arora, R.; Bilmes, J.; Livescu, K. Deep Canonical Correlation Analysis. In Proceedings of the International Conference on International Conference on Machine Learning, Atlanta, GE, USA, 17–19 June 2013. [Google Scholar]
  23. Arthur, T.; Cathy, P.; Vincent, G.; Le, K.A.; Jacques, G.; Vincent, F. Variable Selection for Generalized Canonical Correlation Analysis. Biostatistics 2014, 15, 569–583. [Google Scholar]
  24. Benton, A.; Khayrallah, H.; Gujral, B.; Reisinger, D.; Arora, R. Deep Generalized Canonical Correlation Analysis. In Proceedings of the 4th Workshop on Representation Learning for NLP, Florence, Italy, 2 August 2019; pp. 1–6. [Google Scholar]
  25. Allen-Zhu, Z.; Li, Y. Doubly Accelerated Methods for Faster CCA and Generalized Eigendecomposition. In Proceedings of the 34th International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; pp. 98–106. [Google Scholar]
  26. Arora, R.; Marinov, T.V.; Mianjy, P.; Srebro, N. Stochastic Approximation for Canonical Correlation Analysis. In Proceedings of the Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4775–4784. [Google Scholar]
  27. Bhatia, K.; Pacchiano, A.; Flammarion, N.; Bartlett, P.L.; Jordan, M.I. Gen-Oja: A Simple and Efficient Algorithm for Streaming Generalized Eigenvector Computation. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  28. Yger, F.; Berar, M.; Gasso, G.; Rakotomamonjy, A. Adaptive canonical correlation analysis based on matrix manifolds. In Proceedings of the 29th International Conference on Machine Learning, Edinburgh, UK, 26 June–1 July 2012. [Google Scholar]
  29. Gao, C.; Garber, D.; Srebro, N.; Wang, J.; Wang, W. Stochastic Canonical Correlation Analysis. J. Mach. Learn. Res. 2019, 20, 1–46. [Google Scholar]
  30. Kanatsoulis, C.I.; Fu, X.; Sidiropoulos, N.D.; Hong, M. Structured SUMCOR Multiview Canonical Correlation Analysis for Large-Scale Data. IEEE Trans. Signal Process. 2019, 67, 306–319. [Google Scholar] [CrossRef]
  31. Lu, Y.; Foster, D.P. Large scale canonical correlation analysis with iterative least squares. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 91–99. [Google Scholar]
  32. Chen, Z.H.; Li, X.G.; Yang, L.; Haupt, J.; Zhao, T. On constrained nonconvex stochastic optimization: A case study for generalized eigenvalue decomposition. In Proceedings of the 22nd International Conference on Artifificial Intelligence and Statistics, Naha, Okinawa, Japan, 16–18 April 2019; pp. 916–925. [Google Scholar]
  33. Ma, Z.; Lu, Y.C.; Foster, D.P. Finding linear structure in large datasets with scalable canonical correlation analysis. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 169–178. [Google Scholar]
  34. Kettenring, J.R. Canonical analysis of several sets of variables. Biometrika 1971, 58, 433–451. [Google Scholar] [CrossRef]
  35. Golub, G.H.; Zha, H. The Canonical Correlations of Matrix Pairs and their Numerical Computation. In Linear Algebra for Signal Processing; Bojanczyk, A., Cybenko, G., Eds.; Springer: New York, NY, USA, 1995. [Google Scholar]
  36. Avron, H.; Boutsidis, C.; Toledo, S.; Zouzias, A. Efficient Dimensionality Reduction for Canonical Correlation Analysis. Sci. Comput. 2014, 36, 347–355. [Google Scholar] [CrossRef] [Green Version]
  37. Tropp, J.A. Improved Analysis of the Subsampled Randomized Hadamard Transform. Adv. Data Sci. Adapt. Anal. 2011, 3, 115–126. [Google Scholar] [CrossRef] [Green Version]
  38. Wang, W.R.; Wang, J.L.; Srebro, N. Globally convergent stochastic optimization for canonical correlation analysis. Adv. Neural Inf. Proc. Syst. 2016, 1, 766–774. [Google Scholar]
  39. Ge, R.; Jin, C.; Kakade, S.M.; Netrapalli, P.; Sidford, A. Efficient algorithms for large-scale generalized eigenvector computation and canonical correlation analysis. In Proceedings of the 33rd International Conference on International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 2741–2750. [Google Scholar]
  40. Xu, Z.Q.; Li, P. Towards Practical Alternating Least-Squares for CCA. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  41. Kim, M. Correlation-based incremental visual tracking. Pattern Recognit. 2012, 45, 1050–1060. [Google Scholar] [CrossRef]
  42. Ross, D.A.; Lim, J.; Lin, R.S.; Yang, M.H. Incremental Learning for Robust Visual Tracking. Int. J. Comput. Vis. 2008, 77, 125–141. [Google Scholar] [CrossRef]
  43. Bhatia, K.; Pacchiano, A.; Flammarion, N.; Bartlett, P.L.; Jordan, M.I. Gen-Oja: A Two-time-scale approach for Streaming CCA. arXiv 2018, arXiv:1811.08393. Available online: https://arxiv.org/abs/1811.08393 (accessed on 4 November 2020).
  44. Levey, A.; Lindenbaum, M. Sequential Karhunen-Loeve basis extraction and its application to images. IEEE Trans. Image Process. 2000, 9, 1371–1374. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Pan, J.; Lim, J.; Su, Z.X.; Yang, M.H. L0-Regularized Object Representation for Visual Tracking. In Proceedings of the British Machine Vision Conference, Nottingham, UK, 1–5 September 2014. [Google Scholar]
  46. Adam, A.; Rivlin, E.; Shimshoni, I. Robust Fragments-based Tracking using the Integral Histogram. In Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006. [Google Scholar]
  47. Kwon, J.; Lee, K.M. Visual tracking decomposition. Computer Vision & Pattern Recognition. In Proceedings of the 23rd IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1269–1276. [Google Scholar]
  48. Bao, C.L.; Wu, Y.; Li, H.B.; Ji, H. Real time robust L1 tracker using accelerated proximal gradient approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  49. Babenko, B.; Yang, M.H.; Belongie, S.J. Visual tracking with online Multiple Instance Learning. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 983–990. [Google Scholar]
  50. Wu, Y.; Lim, J.; Yang, M.H. Online Object Tracking: A Benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
Figure 1. Tracking results of several trackers on ten videos. The bounding boxes of several trackers on three representative frames of (a) Basketball, (b) car4, (c) carDark, (d) carscale, (e) david3, (f) deer, (g) faceocc1, (h) faceocc2, (i) fish, and (j) football, respectively.
Figure 1. Tracking results of several trackers on ten videos. The bounding boxes of several trackers on three representative frames of (a) Basketball, (b) car4, (c) carDark, (d) carscale, (e) david3, (f) deer, (g) faceocc1, (h) faceocc2, (i) fish, and (j) football, respectively.
Applsci 10 07827 g001
Figure 2. Tracking results of several trackers on eight videos. The bounding boxes of several trackers on three representative frames of (a) crossing, (b) dog1, (c) jogging-2, (d) jumping, (e) mhyang, (f) mountainBike, (g) singer1, and (h) walking, respectively.
Figure 2. Tracking results of several trackers on eight videos. The bounding boxes of several trackers on three representative frames of (a) crossing, (b) dog1, (c) jogging-2, (d) jumping, (e) mhyang, (f) mountainBike, (g) singer1, and (h) walking, respectively.
Applsci 10 07827 g002
Figure 3. Success rate of seven trackers on different attributes.
Figure 3. Success rate of seven trackers on different attributes.
Applsci 10 07827 g003
Figure 4. Precision rate of seven trackers on different attributes.
Figure 4. Precision rate of seven trackers on different attributes.
Applsci 10 07827 g004
Table 1. The FPS of the compared trackers on eighteen videos.
Table 1. The FPS of the compared trackers on eighteen videos.
ICCAIVTL1APGTLDVTDMILFrag
Basketball31.0122.092.3619.8113.3630.62.74
Car430.822.11.674714.978.2627.737.07
Cardark33.1232.413.2725.2218.9527.166.68
Carscale25.0122.232.5923.7912.9624.663.47
David332.7131.653.0119.6514.1025.453.19
Deer23.721.92.151616.6212.1517.382.68
Faceocc128.8727.314.5617.8513.8027.784.27
Faceocc230.9522.49321.9313.8029.015.32
Fish37.0529.012.2236.1116.4825.025.39
Football31.5322.063.4420.9913.6731.562.87
Crossing34.1227.841.4830.1412.9611.64.89
Dog123.0322.861.6610.038.4418.676.71
Jogging-226.815.30.1826.7119.8230.065.36
Jumping29.768.531.0210.7814.8729.87.41
Mhyang31.347.313.1712.097.7646.33.15
Mountain Bike32.3631.963.0217.9814.2727.933.67
Singer135.7922.492.2923.1914.4934.563.44
Walking33.0832.224.3412.4513.5228.342.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, H.; Sun, D.; Luo, Z. Incremental Canonical Correlation Analysis. Appl. Sci. 2020, 10, 7827. https://doi.org/10.3390/app10217827

AMA Style

Zhao H, Sun D, Luo Z. Incremental Canonical Correlation Analysis. Applied Sciences. 2020; 10(21):7827. https://doi.org/10.3390/app10217827

Chicago/Turabian Style

Zhao, Hongmin, Dongting Sun, and Zhigang Luo. 2020. "Incremental Canonical Correlation Analysis" Applied Sciences 10, no. 21: 7827. https://doi.org/10.3390/app10217827

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop