Next Article in Journal
Mechanisms of High-Grade Serous Carcinogenesis in the Fallopian Tube and Ovary: Current Hypotheses, Etiologic Factors, and Molecular Alterations
Next Article in Special Issue
Inferring Single-Cell 3D Chromosomal Structures Based on the Lennard-Jones Potential
Previous Article in Journal
Weak Electromagnetic Fields Accelerate Fusion of Myoblasts
Previous Article in Special Issue
CBCR: A Curriculum Based Strategy For Chromosome Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Protein Structure Refinement Using Multi-Objective Particle Swarm Optimization with Decomposition Strategy

1
Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, and Key Laboratory of System Control and Information Processing, Ministry of Education of China, Shanghai 200240, China
2
Department of Computer Science, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2021, 22(9), 4408; https://doi.org/10.3390/ijms22094408
Submission received: 30 March 2021 / Revised: 16 April 2021 / Accepted: 20 April 2021 / Published: 23 April 2021
(This article belongs to the Special Issue Protein, RNA, and Genome Structure Prediction)

Abstract

:
Protein structure refinement is a crucial step for more accurate protein structure predictions. Most existing approaches treat it as an energy minimization problem to intuitively improve the quality of initial models by searching for structures with lower energy. Considering that a single energy function could not reflect the accurate energy landscape of all the proteins, our previous AIR 1.0 pipeline uses multiple energy functions to realize a multi-objectives particle swarm optimization-based model refinement. It is expected to provide a general balanced conformation search protocol guided from different energy evaluations. However, AIR 1.0 solves the multi-objective optimization problem as a whole, which could not result in good solution diversity and convergence on some targets. In this study, we report a decomposition-based method AIR 2.0, which is an updated version of AIR, for protein structure refinement. AIR 2.0 decomposes a multi-objective optimization problem into a number of subproblems and optimizes them simultaneously using particle swarm optimization algorithm. The solutions yielded by AIR 2.0 show better convergence and diversity compared to its previous version, which increases the possibilities of digging out better structure conformations. The experimental results on CASP13 refinement benchmark targets and blind tests in CASP 14 demonstrate the efficacy of AIR 2.0.

1. Introduction

The functions of a protein are closely related to its 3D structure, and high-resolution protein structure can increase the understanding of what it does and how it works. In the past decades, dramatic progress has been made in structure determination using wet-lab experimental methods, such as X-ray crystallography, nuclear magnetic resonance (NMR) spectroscopy, and recent electron microscopy techniques [1]. However, these experiments are still expensive and time-consuming [2]. Many popular automated protein structure prediction methods play important complementary roles [3,4,5], such as AlphaFold [6], trRosetta [7], I-TASSER [8], and MULTICOM [9,10]. Especially in recent years, protein structure prediction performance has been largely improved due to the advances in both theoretical and computational studies as demonstrated in recent CASP (Critical Assessment of protein Structure Prediction) assessment, e.g., coevolution analysis-based investigation [11,12,13,14], powerful deep learning computational techniques [15,16,17], etc.
Although remarkable results have been achieved in protein structure prediction, the predicted models still contain inaccurate regions deviating from the native structures [18]. Thus, there have been increasing efforts on improving predicted models via refinement as a following step. Since the 8th competition of Critical Assessment of protein Structure Prediction, the protein structure refinement task has been introduced to evaluate the performance of computational methods for structure refinement by given an initial predicted model [19,20,21]. However, it is a challenging task until now, as it is a blind refinement and on some hard targets, refinement methods degrade their initial models rather than improve them.
One of the common strategies for protein structure refinement is to implement the work pipeline through the combination of energy functions and optimization algorithms [22,23,24,25]. The energy function is designed to describe a protein’s state that is near-native or non-native from its view, which will guide the refinement search to its lower energy state. Considering its importance, a number of molecular mechanics force fields and knowledge-based energy functions have been proposed, i.e., AMBER [26], CHARMM [27], OPLS [28], RWplus [29], DFIRE [30], GOAP [31], and Rosetta [32]. However, it is still difficult to apply a single energy function to exactly describe the states of all proteins due to the large diversities of the protein structures. Each energy function would have its advantages and disadvantages on specific targets, which is a potential reason the performance of the refinement algorithms often varies with the targets in the CASP experiments.
In addition to the energy functions, the optimization algorithms are also crucial in protein structure refinement, which are designed to search for the lowest-energy structure conformation. Popular optimization algorithms include Molecular Dynamics (MD) simulation [33] and Monte Carlo (MC) simulation [34]. It is still very challenging to achieve consistent refinement over initial models, and one potential reason is that most existing approaches are conducted based on a single energy function.
Motivated by those observations, we have developed one multi-objective-based refinement method called AIR [35] to alleviate the potential bias caused by minimizing only one energy function. The AIR is a multi-objective particle swarm optimization (PSO)-based protocol [36], where each structure is treated as a particle. The quality of the particles in each iteration is evaluated by three energy functions based on dominance relations [37], and the non-dominated particles are put into a set called Pareto set ( P S ) [38], which is used to select the final refined structures.
However, the dominance-based AIR has no direct control over the movement of each particle in the swarm and no suitable mechanism to maintain the diversity of Pareto front ( P F ) [39]. The loss of diversity may deteriorate the advantage of multi-objective optimization. Moreover, the crucial parameter G b e s t in PSO is difficult to choose, since there are many non-dominated candidates in the P S . Using Pareto dominance alone would deteriorate the selection pressure toward the P F and slow down the searching process [40], since the update of another important parameter P b e s t only needs to reduce one energy function.
To solve the above problems, we present a decomposition-based approach AIR 2.0, which is an updated version of AIR 1.0 to further increase the conformation optimization capability. In AIR 2.0, each particle is associated with a unique subproblem defined by a weight vector, which is different from the protocol of AIR 1.0 that solves a multi-objective optimization problem as a whole. Thus, the diversity is accordingly improved, since each particle is moving toward P F in its own direction. In addition, the P b e s t and G b e s t of each particle have a determined choice according to its own subproblem, helping avoid oscillation in the searching process. The benchmark tests of CASP13 refinement targets and blind tests in CASP14 demonstrate the efficacy of the new updated version of AIR refinement pipeline.

2. Experiments and Results

We have evaluated AIR 2.0 pipeline on the refinement targets of CASP13 and CASP14. To demonstrate the advantage of AIR 2.0, we compare it with state-of-the-art methods, including its previous version AIR 1.0 and other state-of-the-art refinement methods in CASP13 such as FEIGLAB [18], BAKER [24], and Zhang-Refinement [41]. Global distance test total score (GDT-TS) [42], template modeling score (TM-score) [43], and root mean square deviation (RMSD) are the metrics for evaluating the effectiveness of AIR 2.0.
For each target, the number of divisions H in (10) (see Methods) is set to 10 according to our local tests, resulting in N = ( H + M 1 M 1 ) = ( 10 + 3 1 3 1 ) = 66 weight vectors or subproblems and the same number of particles. M = 3 is the number of objectives. The single initial model provided by CASP is taken as input, and another 65 models are generated by applying random perturbations to the initial model. The neighborhood size T is set to 8 according to [44], and the maximum number of iterations is set to 3000 as AIR 1.0. We set S = 20,000 in (9) (see Methods) to get a stable result and output the top five ranked solutions.

2.1. Effectiveness of AIR 2.0 on CASP13

We test AIR 2.0 on the 29 CASP13 refinement targets (two cancelled targets were excluded), and the results are summarized in Figure 1. We compare the best model and Model 1 with the initial model. The best model achieves consistent improvements over the initial model and almost all targets are to a certain degree refined. The average gains in the quality of the best model are +1.98 in GDT-TS, +0.014 in TM-score, and −0.18 Å in RMSD. Compared to Model 1, the average improvement in GDT-TS is 1.22, and 82% of the targets (24 out of 29 targets) are refined. In terms of TM-score and RMSD, the average improvements are +0.0076 and −0.0752 Å with 72% (21 out of 29 targets) and 69% (20 out of 29 targets) being refined, respectively.
It is observed that the targets with a medium quality are more likely to be refined (Figure 2). Specifically, AIR 2.0 improves those targets with the following quality: (1) initial GDT-TS is between 60 and 80, (2) initial RMSD is between 2 and 5 Å, and (3) initial TM-score is between 0.65 and 0.8. The potential reason is that high quality models leave a few spaces to refine, while the relatively bad models might be trapped in a deep local minimum caused by a rough energy landscape.

2.2. AIR 2.0 Is Superior to AIR 1.0

We compare the updated method AIR 2.0 with our previous AIR 1.0. The CASP13 results of AIR 2.0 and AIR 1.0 are summarized in Table 1. Figure 3 illustrates the GDT-TS of all targets refined by AIR 2.0 and 1.0 based on the best model and Model 1. The results indicate that AIR 2.0 achieves better or comparable performance over AIR 1.0. Compared to AIR 1.0, AIR 2.0 obtains a better quality in 21 out of 29 targets for the best model and 24 of 29 targets for Model 1.
For AIR 1.0, a refinement of hard targets in CASP13 often obtained model degradation rather than improvement, such as R0949, R0977D2, R0996D4, and R1016. However, AIR 2.0 achieves promising refinement results on these hard targets. The potential reason is due to the diversity of P S introduced by the decomposition strategy. In the case of R0981D5, the non-dominance solutions in the final P S of AIR 2.0 and 1.0 are plotted in Figure 4b. We can see that AIR 2.0 finds more non-dominated solutions than AIR 1.0, and these solutions are distributed with a high diversity. Moreover, as shown in Figure 4a,c,d, the solutions obtained by AIR 2.0 completely dominate those obtained by AIR 1.0, indicating a more convergence toward the true P F . Thus, the overall quality of the solutions obtained by AIR 2.0 is higher than those of AIR 1.0, which is beneficial for the selection of high quality Model 1.
The dominance-based method AIR 1.0 drives the whole population toward the P F without direct control over the movement of each individual in the population. Thus, AIR 1.0 prefers the regions that are easy to access and does not sufficiently account for the diversity. As a result, the solutions obtained by AIR 1.0 are only distributed in a small area. Moreover, due to the lack of stable guidance on P b e s t and G b e s t (see Methods), the searching process of each particle would be difficult. However, the decomposition strategy in AIR 2.0 assigns a single objective optimization subproblem for each particle. In this way, each particle has an exact update direction or a clear target position in P F , which results in better diversity and convergence features. This is the potential reason why AIR 2.0 outperforms AIR 1.0 in most targets.

2.3. Comparison with Other State-of-the-Art Refinement Methods

The test data consist of those targets in which each group performs the best on CASP13 in order to highlight the characteristics of each method. The refined models of BAKER, FEIGLAB, and Zhang are available at the CASP official website. As shown in Table 2, AIR 2.0 yields promising improvement over initial models. For the target R0949, R0979, and R0989D1, AIR 2.0 is the only method that achieves improvement rather than degradation over initial models and the GDT-TS gains is 0.59, 3.53, and 0.37, respectively. However, for the targets such as R0968s1, R0974s1, and R0986s1, the GDT-TS gains obtained by BAKER or FEIGLAB are larger than AIR 2.0. The potential reason is probably that AIR 2.0 uses multiple energy functions that constrain each other, resulting in a limited deviation from initial models. These results also highlight that the protein structure map is huge, and it is very hard to achieve a general better refinement algorithm on all proteins. For hard targets, AIR 2.0 would be reliable, since we extend the one-dimension optimization to a new three-dimension space optimization, partially alleviating the bias caused by using only one energy function.

2.4. Blind Test in CASP14

We also test our method in the recent CASP14 blind test. Overall, our AIR ranks the ninth among 37 groups in the competition according to SUM Zscore > −2.0 (including the reference group named STARTING MODEL). There are 51 refinement targets in total, where two targets were cancelled during the competition and five targets do not have a native structure for reference. The results of AIR on the remaining 44 targets are summarized in Figure 5 (for more details, please refer to the CASP14 website). The average gains in the quality of the best model among Model 1–5 are +0.36 in GDT-TS, which is slightly lower than CASP13. The main reason is that our solution ranking method performed relatively poorly on these targets, and we found locally that a number of better structures were not selected as the top five submission models. This is also one of our future efforts to further improve the AIR program. However, there are also some successful cases. For example, AIR ranks the first on the target R1042v2 among the models submitted by 31 groups (https://predictioncenter.org/casp14/results.cgi?view=tables&target=R1042v2&model=1&groups_id=, accessed on 21 April 2021). Considering the Model 1 submission models, the AIR approach is the only one that successfully achieves improvement rather than degradation over the initial model. Considering the best submission models, our predictions for the target R1029 are among the most accurate in all submissions (https://predictioncenter.org/casp14/results.cgi?view=tables&target=R1029&model=all&groups_id=, accessed on 21 April 2021).
The success of these two cases indicates the potential advantages of multi-objective optimization and the PSO algorithm that can efficiently explore the high-dimensional energy landscape to get a reliable refined model. On the other hand, the performance of AIR in CASP14 also indicates that there is still room for improvement of our algorithm. For instance, on the target R1042v2, the improvement is still limited to a moderate level. For target R1029, our Model 2 submission is better than our Model 1, implying that we still need to investigate how to rank the final solution. In our AIR’s future development, we will go on to find a better mechanism that could guide the search process to achieve significant improvement and a new method to accurately score all the candidate solutions.

3. Discussion

3.1. The Importance of the Diversity on AIR 2.0

In AIR 1.0, we have shown that multi-objective optimization is a promising way to improve protein structure refinement. The two goals of the multi-objective optimization are: (1) a set of solutions as close as possible to the P F ; (2) a broadly distributed set of solutions that cover the entire P F [45]. The two goals are also referred to convergence and diversity. In the field of protein structure refinement, the diversity of the solutions is important. When given a model to be refined, we have no idea which direction or what combination of multiple energy functions is feasible for improvement due to the diversity of protein structures. In order to improve the initial model, AIR 2.0 tries different directions to obtain a well-distributed P F that covers all potential solutions. As mentioned before, dominance-based method AIR 1.0 prefers the regions that are easy to access, resulting in insufficient diversity, which may lose the solution diversity.
To give a more intuitive understanding of the conformation solution diversity, Figure 6 shows a comparison between AIR 1.0 and AIR 2.0 on the target R0949 from CASP13 whose P F is irregular, consisting of at least two parts. The solutions of AIR 1.0 cover only one part, and the GDT-TS of the Model 1 is 62.98, which is a degradation of the initial model with a GDT-TS of 64.53. However, Model 1 of AIR 2.0 yields a GDT-TS of 65.12 on the other part of the P F , demonstrating improvement over the initial model. Therefore, if we do not take the diversity carefully, the possibility of improvement would decrease.

3.2. The Influence of Hyperparameters on AIR 2.0

The neighborhood size T (see Methods) is a major control parameter in AIR 2.0 since the solutions in the neighborhood of a subproblem can be used to guide the searching process. In a sense, each subproblem with its neighborhood is regarded as a swarm. To test the influence of T , we perform some experiments on those targets with different size. The results are summarized in Table 3. When T = 3 , the neighborhood is too small, and the particles in a swarm are similar, resulting in the inability to explore the new area and achieve a good result. It should also be noted that AIR 2.0 performs similarly with T larger than 8. That is because the g b e s t for each subproblem depends on only a certain number of neighbors, while others play a small role. Moreover, a large T will increase the computational burden and might undermine the diversity of solutions. Thus, we set T = 8 to make a tradeoff between the performance and running time.
The penalty value θ (see Methods) in the PBI decomposition approach is another important parameter [46]. In this study, we adopt an adaptive penalty scheme (APS) [47] that linearly increases θ with the number of generations from 5 to 20. At the early stage, a small θ is beneficial for convergence toward PF, and the value of θ is gradually increased to promote the diversity of solutions. For the number of iterations M a x I t and the number of particles N , generally, the larger the two numbers are, the better the performance is. However, large values of these two parameters will increase the time cost. To make a tradeoff, we finally set M a x I t = 3000 and N = 66 .

4. Methods

4.1. Overview of Refinement Pipeline AIR 2.0

As shown in Figure 7, AIR 2.0 consists of three main steps. The first step is swarm initialization, which generates multiple particles. If a single initial model is used as the input, other particle models can be generated by applying perturbations to the initial one. In total, an initial swarm of N particles is obtained. In the second step, each particle is associated with a unique weight vector generated by the simplex-lattice design method [48]. Then, the main loop of optimization is performed, where Rosetta, RWplus, and CHARMM are selected as three fitness functions. Each particle updates the position according to its own subproblem formulated by the weight vector. In each iteration, the non-dominated solutions in the whole population are added to the Pareto set. In the third step, all solutions in the Pareto set are ranked using the expected utility rule [49] and the top five of them are selected as the final refined structures.

4.2. Representations of Protein Conformations

Mathematically, protein conformation could be represented by the Cartesian coordinates of the atoms or internal coordinates (bond lengths and angles) [50]. The former is suitable for describing physical force fields, and the latter is a better representation to describe bonded interactions as well as certain kinds of experimental information [51]. In AIR 2.0, we use both coordinate systems.
During the sampling stage, the protein backbone is represented by a list of main-chain torsion angles using internal coordinates:
C = [ ϕ 1 , ψ 1 , ω 1 , , ϕ L , ψ L ]
where L stands for the protein length. We further use the Denavit and Hartenberg (DH) method [52] to convert internal coordinates to corresponding Cartesian coordinates, since certain energy functions, such as the Rosetta, explicitly encode Cartesian energy terms. This conversion goes back and forth until the end of the pipeline.

4.3. Multi-Objective Optimization

Similar to its previous version, the AIR 2.0 uses three energy functions to perform conformation search in a 3D energy space composed by Rosetta, RWplus, and CHARMM. It is crucial to select accurate force fields for protein structure refinement. There are roughly two types of force fields in the community. One is physics-based force fields that are designed on the basis of all kinds of interactions at the atomic and molecular levels. The other is knowledge-based energy function deduced from diverse sets of known protein structures. Each type of force field has it its merits and drawbacks. To take advantage of both types of force fields, we choose one popular physics-based force field CHARMM and one typical knowledge-based energy function RWplus. Rosetta energy function could be classified into both types and is widely used in protein structure prediction and refinement for its good performance. Therefore, we use it as a complementary part for the other two force fields. This will formulate the protein structure refinement as a multi-objective optimization (MOP) problem as follows:
m i n i m i z e   F ( C ) = ( f R o s e t t a ( C ) , f R W p l u s ( C ) , f C H A R M M ( C ) ) T s u b j e c t   t o   C Ω    
where C is the conformation of a protein and Ω is the overall conformational space. f R o s e t t a ( C ) , f R W p l u s ( C ) , and f C H A R M M ( C ) are three energy values in terms of C .
Due to potential conflicts among multiple objectives, usually, one single solution (conformation) cannot optimize all objectives simultaneously. Instead, a set of optimal solutions representing the trade-offs among different objectives could be obtained. A dominance relation between different solutions is often used to suggest the acceptance of current conformations.
Let C i , C j Ω ; we say that C i   d o m i n a t e s   C j   (denoted as C i < C j ) if and only if k = 1 , 2 , 3 , f k ( C i ) f k ( C j ) and F ( C i ) F ( C j ) , where f 1 ,   f 2 , and f 3 correspond to f R o s e t t a ( C ) ,   f R W p l u s ( C ) , and f C H A R M M ( C ) respectively. If C * Ω and there is no other solution in Ω that dominates C * , then C * is considered as a Pareto optimal solution. The Pareto set ( P S ) is defined as:
P S = { C Ω | C   i s   a   P a r e t o   o p t i m a l   s o l u t i o n } .
The energy map of all Pareto optimal solutions in P S is called a Pareto front ( P F ) [53] and can be described as:
P F = { F ( C ) = ( f R o s e t t a ( C ) , f R W p l u s ( C ) , f C H A R M M ( C ) ) T | C P S } .
The goal of multi-objective optimization is to obtain widely distributed Pareto optimal solutions that are as close to true P F as possible.

4.4. Decomposition Approach in Multi-Objective Optimization

Our original protocol AIR 1.0 solves the multi-objective optimization based on Pareto dominance [37]. It mainly evaluates each solution by its Pareto dominance relations to other solutions and aims to drive the population toward the P F as a whole. However, the movement of each particle in the population and the distribution of computational effort over different ranges of the P F can be further investigated. Otherwise, the whole population would prefer to the regions that are easily accessible and cannot maintain the diversity of the solutions.
Generally speaking, a Pareto optimal solution for the multi-objective optimization problem can be seen as the optimal solution of a scalar optimization subproblem whose objective is an aggregation function of all the individual objectives ( f R o s e t t a ,   f R W p l u s , f C H A R M M ) [44]. Thus, a multi-objective optimization problem can be decomposed into a number of optimization subproblems, and each subproblem is distinguished by one unique weight vector. Then, Pareto solutions could be achieved by minimizing such subproblems. There exists several methods to construct the aggregation function [54] for each subproblem with a weight vector, such as weight sum approach [44], Tchebycheff approach [55], penalty-based boundary intersection (PBI) [44], etc. Here, AIR 2.0 uses the PBI approach to construct the aggregation function for each subproblem. Formally, an optimization subproblem in AIR 2.0 can be stated as:
m i n i m i z e   g p b i ( C | λ i , z * ) = d 1 + θ d 2 s u b j e c t   t o   C Ω
where
d 1 = ( F ( C ) z * ) T λ i λ
d 2 = F ( C ) ( z * + d 1 λ i λ i )
where C is a candidate solution (conformation) that belongs to overall conformational space. F ( C ) = ( f R o s e t t a ( C ) , f R W p l u s ( C ) , f C H A R M M ( C ) ) T consists of three components from Rosetta, RWplus, and CHARMM. λ i = ( λ i 1 , λ i 2 , λ i 3 ) T is the weight vector of the i th subproblem satisfying λ i j 0 and j = 1 3 λ i j = 1 . z * = ( z R o s e t t a * , z R W p l u s , z C H A R M M * ) T is the ideal objective vector with z k * min C Ω f k ( C ) ,   k { R o s e t t a , R W p l u s , C H A R M M } ,   θ is a user-defined penalty parameter. d 1 is the distance between the ideal objective vector z * and the solution F ( C ) , d 2 is the direction error between λ i and F(C). The PBI approach tries to minimize both d 1 and d 2 , and their relative importance is control by θ .
Figure 8 presents a simple example to illustrate the PBI approach given the weight vector λ i = ( 0.33 ,   0.33 , 0.33 ) T . F(C) and z * are denoted as two points in the energy map. The orange plane is P F , and d 1 , d 2 are marked in the figure. It is clear that the intersection of the weight vector and P F , which is marked by a black point, is the optimal solution of the subproblem defined by PBI with λ i   .
Thus, the optimal solution to (5) is a Pareto optimal solution to (2). We use λ i to emphasize that (5) is the i th subproblem defined by a weight vector. In order to obtain a set of different Pareto optimal solutions, we can use different weight vectors. A natural idea comes that if we have a large number of uniformly distributed weight vectors, we could get a set of Pareto optimal solutions that approximates P F very well.

4.5. Particle Swarm Optimization

Particle swarm optimization (PSO) [56] is a meta-heuristic algorithm simulating the behaviors of groups of birds and fishes. It solves a problem by iteratively improving candidate solutions with the information coming from their population. Each candidate solution not only has its own exploration behavior, but its trajectory is also affected by other solutions in the population. In PSO, every individual in the population is called a particle, and swarm is another name for population. In AIR 2.0, each particle in the swarm represents a candidate conformation in the overall conformational space. A particle is characterized by its position and velocity, where the position is the conformation of a protein represented by (1) and the velocity represents the change of torsion angles. The particle uses the position of the selected global leader and its own personal movement trajectory to update the velocity and position values using (8) and (9).
v i t + 1 = w v i t + c 1 r 1 ( P b e s t i t C i t ) + c 2 r 2 ( G b e s t i t C i t )
C i t + 1 = C i t + v i t + 1
where v i t is the velocity of the i th particle in the t th generation, C i t is the new conformation of the i th particle in the t th generation, and w is the inertia weight.
According to our previous study [35], we set w to 1.3 at the beginning; it linearly decreases to 0.7 as the number of iterations increases.   c 1 and   c 2 are two learning coefficients that are both set to be 2 [57].   r 1 , r 2 [ 0 , 1 ] are uniformly distributed random variables. P b e s t i t is the best conformation that the i th particle has ever been until the t th generation. Similarly, G b e s t i t is the best conformation that the whole swarm has ever met until the t th generation. Each time the conformation updates, the non-dominated ones are added into P S .
For only one objective, each solution can be ranked according to the objective. Thus, both P b e s t i t and G b e s t i t have a determined choice. However, for multiple objectives, there are always many non-dominated solutions that are equally good under the concept of the dominance. Thus, it is difficult to choose P b e s t i t and G b e s t i t to lead the searching process. In our previous protocol of AIR 1.0, P b e s t i t is updated when any one of the three energy functions decreases, and G b e s t i t is randomly selected from current P S . This will severely deteriorate the selection pressure toward the PF and considerably slows down the searching process due to ambiguous search direction.

4.6. Obtaining Pareto Optimal Set with Multi-Objective Particle Swarm Optimization Based on Decomposition Strategy

Due to the diversity of protein structures, we use multiple energy functions as multi-objectives to alleviate the bias problem caused by minimizing one single energy function. Given a particular protein, which energy function or what combination of these energy functions is appropriate for a particular protein is still unknown. Each candidate solution on P F represents a potential optimization direction. Thus, the diversity of solutions on the P F is of importance for multi-objective optimizations in protein structure refinement. In order to make good use of three energy functions, we need to find as many Pareto optimal solutions as possible and maximize the distribution of solutions in the P F . However, using Pareto dominance alone could discourage the diversity of solutions, since it has no direct control over the movement of each individual in its population and no good mechanism to control the distribution of the computational effort over different ranges of the P F . As a result, the whole population is updated in a random direction and prefers those regions that are easily accessible. Finally, the solutions will end up in a small range of P F , resulting in the loss of the diversity.
In order to overcome the above shortcomings, AIR 2.0 uses a decomposition strategy to define a single objective optimization subproblem for each particle. A Pareto optimal solution to an MOP could be an optimal solution of a scalar optimization subproblem, in which the objective is an aggregation of all the objectives in AIR 2.0. In this way, each particle has an exact updating direction and increasing evolutionary pressure, which is beneficial to the convergence. In addition, the diversity is inherently guaranteed since each particle is moving toward P F in its own direction. The general framework is as follows.
At the beginning, a set of weight vectors { λ 1 , , λ N } ( N is the number of particles) are generated using the canonical simplex-lattice design method [48], whose weight vectors are sampled from a unit simplex.
{ λ i = ( λ i 1 , λ i 2 , λ i 3 )                      λ i j { 0 H , 1 H , , H H } , j = 1 3 λ i j = 1
where i = 1 , , N is the index of uniformly distributed weight vector. λ i has three components corresponding to three energy functions, Rosetta, RWplus, and CAHRMM. H > 0 is the number of divisions along each objective coordinate. In total, there are N = ( H + M 1 M 1 ) = ( H + 3 1 3 1 ) different weight vectors for M = 3 objectives. Then, each particle is associated with a different weight vector, which defines a unique subproblem. Solving these subproblems is equivalent to solving the original multi-objective optimization problem.
In AIR 1.0, the velocity and position of a particle are updated using the information from its individual best P b e s t i t and the global best G b e s t i t . However, it is difficult to select a suitable one, since multiple objectives result in a large number of equally good non-dominated solutions. Now with the decomposition strategy, P b e s t i t is obvious for a particle with a weight vector λ i using the aggregation function g P B I ( C | λ i , z * ) . For G b e s t i t , there is a small difference, since each particle corresponds to a different subproblem. However, if two weight vectors λ i and λ j are close enough, the optimal solutions to both two subproblems, g P B I ( C | λ i , z * ) and g P B I ( C | λ j , z * ) , will also be similar. Therefore, the information from the searching process of λi is useful to λ j and vice versa. According to this observation, a neighborhood of the weight vector λ i   is defined as a set of weight vectors { λ i 1 , λ i 2 , λ i T } that are closest to λ i , where T is the size of the neighborhood. Correspondingly, the neighborhood of the i th particle is composed of those particles whose weight vectors are in the neighborhood of λ i . With the notion of neighborhood, G b e s t i t is defined as the best position in the neighborhood of the i th particle during t generations. Then, we could use the particle swarm optimization algorithm to optimize those subproblems simultaneously and finally obtain a Pareto optimal set. The pseudocode of the main framework for AIR 2.0 is summarized in Algorithm 1.
Algorithm 1 Main Framework of AIR 2.0
Input: Initial model C 0 , the maximum number of iterations MaxIT, the number of particles N.
Output: Pareto set P S .
   /*Initialization*/
  • Generate weight vectors { λ 1 , , λ N } using the simplex lattice method.
  • Create initial population { C 1 0 , , C N t } by perturbing C 0 and assign weight vectors to each particle individually.
  • Compute the Euclidean distances between any two weight vectors. For each particle C i 0 , i = 1 , , N , set B ( i ) = { i 1 , .. i T } , where λ i 1 , , λ i T are the T closest weight vectors to λ i .
  • Initialize ideal objective vector z * , set the initial velocity randomly, p b e s t i 0 = g b e s t i 0 = C i 0 , add initial non-dominated particles into P S .
       /*Main Loop*/
  • while t < M a x I t do:
  • for i = 1 , , N do:
  • Update the position of the C i t using PSO Formulas (8) and (9).
  • Update z *
  • if g ( C i   t | λ i ) < g ( p b e s t i t | λ i ) then
  • p b e s t i t = C i   t
  • end if
  • for each j B ( i ) do:
  • if g ( C i   t | λ j ) < g ( g b e s t j t | λ j ) then
  • g b e s t j t = C i t
  • end if
  • end for
  • Remove all the vectors dominated by C i t from P S .
  • Add C i t into P S if no vectors in P S dominate it.
  • end for
  • end while

4.7. Model Selection

After enough iterations, there are plenty of non-dominated solutions or candidate models in the PS. To obtain the final refined structures, we need to assess and rank the generated models. Many methods for estimation of model accuracy have been described such as MULTICOM_CLUSTER [10], Pcons [58], PRESCO [59], DeepAccNet [60], and ProQ3D [61]. Here, for a quick ranking purpose, we use a widely used knee-based ranking method [49] in a multi-objective optimization problem to rank those models. Since the three energy functions are treated equally, AIR 2.0 has no preference to any regions of the P F . However, there will be some special solutions called ‘knee’ in the P F . In such ‘knees’ solutions, a small improvement in one objective will cause large depravation in other objectives. Thus, three objectives reach a balance that all objectives are relatively optimal and no objective can decrease further without seriously increasing other objectives.
To obtain these ‘knee’ solutions, we adopt a utility-based method similar to the AIR 1.0, which uses the expected marginal utility to measure the importance of the solutions in the P S :
U C , w = w 1 f 1 ( C ) + w 2 f 2 ( C ) + w 3 f 3 ( C ) ) s . t .      w 1 + w 2 + w 3 = 1   a n d   w 1 , w 2 , w 3 0
where C is the non-dominated solution in the P S and w 1 , w 2 , w 3 are the weight coefficients. The expected margin utility is approximated by random sampling of w i . For each conformation, we obtain a large number of utility values { U C , w i , i = 1 , S } by using different combinations of weight coefficients. The expected margin utility could be approximated by the average of these sample values:
E ( U C , w ) = 1 S i = 1 S U C , w i .
Then, we can rank the solutions in P S according to individual expected marginal utility and output the top-ranking solutions.

5. Conclusions and Future Direction

In this study, we developed a decomposition-based method AIR 2.0 for protein structure refinement. AIR 2.0 employs a decomposition strategy that divides the multi-objective optimization into a set of subproblems and optimizes them in a collaborative manner. The performance on CASP 13 refinement targets and a blind test on CASP 14 shows that AIR 2.0 is capable of achieving promising results. In the future, we will further improve the AIR refinement protocol to use deep learning methods to design new energy functions that could guide the search process and identify those local structure regions need to be refined. With this information, we could reduce the searching space largely and make the sampling process more efficient. Moreover, the new energy function could be used as the final model selection criterion to rank the models in Pareto set, which may bridge the gap between Model 1 and the best model.

Author Contributions

Conceptualization, C.-P.Z. and H.-B.S.; Data curation, D.W.; Formal analysis, C.-P.Z., D.W. and H.-B.S.; Funding acquisition, H.-B.S.; Investigation, C.-P.Z.; Methodology, C.-P.Z., D.W., X.P. and H.-B.S.; Project administration, H.-B.S.; Resources, H.-B.S.; Software, C.-P.Z.; Supervision, H.-B.S.; Validation, C.-P.Z., D.W. and H.-B.S.; Visualization, C.-P.Z.; Writing—Original draft, C.-P.Z.; Writing—Review & editing, X.P. and H.-B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61725302, 62073219, 61903248, 92059206), and the Science and Technology Commission of Shanghai Municipality (No. 20S11902100).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data about our method could be obtained through the implementation above. The data about other methods is available on CASP official website.

Conflicts of Interest

The authors declare no conflict of interest.

Availability and Implementation

AIR 2.0 is available online: www.csbio.sjtu.edu.cn/bioinf/AIR/ (accessed on 21 April 2021).

References

  1. Schwede, T.; Peitsch, M.C. Computational Structural Biology: Methods and Applications; World Scientific: Singapore, 2008. [Google Scholar]
  2. Dill, K.A.; MacCallum, J.L. The protein-folding problem, 50 years on. Science 2012, 338, 1042–1046. [Google Scholar] [CrossRef] [Green Version]
  3. Kihara, D.; Lu, H.; Kolinski, A.; Skolnick, J. TOUCHSTONE: An ab initio protein structure prediction method that uses threading-based tertiary restraints. Proc. Natl. Acad. Sci. USA 2001, 98, 10125–10130. [Google Scholar] [CrossRef] [Green Version]
  4. Leaver-Fay, A.; Tyka, M.; Lewis, S.M.; Lange, O.F.; Thompson, J.; Jacak, R.; Kaufman, K.; Renfrew, P.D.; Smith, C.A.; Sheffler, W.; et al. ROSETTA3: An object-oriented software suite for the simulation and design of macromolecules. Methods Enzymol. 2011, 487, 545–574. [Google Scholar]
  5. Xu, D.; Zhang, Y. Ab initio protein structure assembly using continuous structure fragments and optimized knowledge-based force field. Proteins Struct. Funct. Bioinform. 2012, 80, 1715–1735. [Google Scholar] [CrossRef] [Green Version]
  6. Senior, A.W.; Evans, R.; Jumper, J.; Kirkpatrick, J.; Sifre, L.; Green, T.; Qin, C.; Zidek, A.; Nelson, A.W.R.; Bridgland, A.; et al. Improved protein structure prediction using potentials from deep learning. Nature 2020, 577, 706–710. [Google Scholar] [CrossRef]
  7. Yang, J.; Anishchenko, I.; Park, H.; Peng, Z.; Ovchinnikov, S.; Baker, D. Improved protein structure prediction using predicted interresidue orientations. Proc. Natl. Acad. Sci. USA 2020, 117, 1496–1503. [Google Scholar] [CrossRef]
  8. Yang, J.; Yan, R.; Roy, A.; Xu, D.; Zhang, Y. The I-TASSER suite: Protein structure and function prediction. Nat. Methods 2014, 12, 7–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Renzhi, C.; Debswapna, B.; Badri, A.; Jilong, L.; Jianlin, C. Massive integration of diverse protein quality assessment methods to improve template based modeling in CASP11. Proteins Struct. Funct. Bioinform. 2015, 84, 247–259. [Google Scholar]
  10. Hou, J.; Wu, T.; Cao, R.; Cheng, J. Protein tertiary structure modeling driven by deep learning and contact distance prediction in CASP13. Proteins Struct. Funct. Bioinform. 2019, 87, 1165–1178. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Kamisetty, H.; Ovchinnikov, S.; Baker, D. Assessing the utility of coevolution-based residue-residue contact predictions in a sequence- and structure-rich era. Proc. Natl. Acad. Sci. USA 2013, 110, 15674–15679. [Google Scholar] [CrossRef] [Green Version]
  12. Wu, T.; Guo, Z.; Hou, J.; Cheng, J. DeepDist: Real-value inter-residue distance prediction with deep residual convolutional network. BMC Bioinform. 2021, 22, 1–17. [Google Scholar] [CrossRef]
  13. Adhikari, B.; Cheng, J. CONFOLD2: Improved contact-driven ab initio protein structure modeling. BMC Bioinform. 2018, 19, 22. [Google Scholar] [CrossRef] [Green Version]
  14. Adhikari, B.; Hou, J.; Cheng, J. Protein contact prediction by integrating deep multiple sequence alignments, coevolution and machine learning. Proteins Struct. Funct. Bioinform. 2018, 86, 84–96. [Google Scholar] [CrossRef] [PubMed]
  15. Kandathil, S.M.; Greener, J.G.; Jones, D.T. Recent developments in deep learning applied to protein structure prediction. Proteins 2019, 87, 1179–1189. [Google Scholar] [CrossRef] [PubMed]
  16. Lee, G.R.; Won, J.; Heo, L.; Seok, C. GalaxyRefine2: Simultaneous refinement of inaccurate local regions and overall protein structure. Nucleic Acids Res. 2019, 47, W451–W455. [Google Scholar] [CrossRef] [PubMed]
  17. Hou, J.; Adhikari, B.; Cheng, J. DeepSF: Deep convolutional neural network for mapping protein sequences to folds. Bioinformatics 2018, 34, 1295–1303. [Google Scholar] [CrossRef] [PubMed]
  18. Heo, L.; Arbour, C.F.; Feig, M. Driven to near-experimental accuracy by refinement via molecular dynamics simulations. Proteins 2019, 87, 1263–1275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Hovan, L.; Oleinikovas, V.; Yalinca, H.; Kryshtafovych, A.; Saladino, G.; Gervasio, F.L. Assessment of the model refinement category in CASP12. Proteins 2018, 86 (Suppl. 1), 152–167. [Google Scholar] [CrossRef] [Green Version]
  20. Modi, V.; Dunbrack, R.L., Jr. Assessment of refinement of template-based models in CASP11. Proteins 2016, 84 (Suppl. 1), 260–281. [Google Scholar] [CrossRef] [Green Version]
  21. Read, R.J.; Sammito, M.D.; Kryshtafovych, A.; Croll, T.I. Evaluation of model refinement in CASP13. Proteins 2019, 87, 1249–1262. [Google Scholar] [CrossRef] [Green Version]
  22. Heo, L.; Feig, M. Experimental accuracy in protein structure refinement via molecular dynamics simulations. Proc. Natl. Acad. Sci. USA 2018, 115, 13276–13281. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Heo, L.; Park, H.; Seok, C. GalaxyRefine: Protein structure refinement driven by side-chain repacking. Nucleic Acids Res. 2013, 41, W384–W388. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Park, H.; Lee, G.R.; Kim, D.E.; Anishchenko, I.; Cong, Q.; Baker, D. High-accuracy refinement using Rosetta in CASP13. Proteins 2019, 87, 1276–1282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Terashi, G.; Kihara, D. Protein structure model refinement in CASP12 using short and long molecular dynamics simulations in implicit solvent. Proteins Struct. Funct. Bioinform. 2018, 86, 189–201. [Google Scholar] [CrossRef] [PubMed]
  26. Tian, C.; Kasavajhala, K.; Belfon, K.A.A.; Raguette, L.; Huang, H.; Migues, A.N.; Bickel, J.; Wang, Y.; Pincay, J.; Wu, Q.; et al. ff19SB: Amino-Acid-Specific Protein Backbone Parameters Trained against Quantum Mechanics Energy Surfaces in Solution. J. Chem. Theory Comput. 2020, 16, 528–552. [Google Scholar] [CrossRef] [PubMed]
  27. Huang, J.; Rauscher, S.; Nawrocki, G.; Ran, T.; Feig, M.; de Groot, B.L.; Grubmuller, H.; MacKerell, A.D., Jr. CHARMM36m: An improved force field for folded and intrinsically disordered proteins. Nat. Methods 2017, 14, 71–73. [Google Scholar] [CrossRef] [Green Version]
  28. Jorgensen, W.L.; Maxwell, D.S. Development and testing of the OPLS all-atom force field on conformational energetics. J. Am. Chem. Soc. 1996, 118, 11225–11236. [Google Scholar] [CrossRef]
  29. Zhang, J.; Zhang, Y. A Novel Side-Chain Orientation Dependent Potential Derived from Random-Walk Reference State for Protein Fold Selection and Structure Prediction. PLoS ONE 2010, 5, e15386. [Google Scholar] [CrossRef] [Green Version]
  30. Zhou, H.; Zhou, Y. Distance-scaled, finite ideal-gas reference state improves structure-derived potentials of mean force for structure selection and stability prediction. Protein Sci. 2002, 11, 2714–2726. [Google Scholar] [CrossRef] [Green Version]
  31. Zhou, H.; Skolnick, J. GOAP: A Generalized Orientation-Dependent, All-Atom Statistical Potential for Protein Structure Prediction. Biophys. J. 2011, 101, 2043–2052. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Alford, R.F.; Leaver-Fay, A.; Jeliazkov, J.R.; O’Meara, M.J.; DiMaio, F.P.; Park, H.; Shapovalov, M.V.; Renfrew, P.D.; Mulligan, V.K.; Kappel, K.; et al. The Rosetta All-Atom Energy Function for Macromolecular Modeling and Design. J. Chem. Theory Comput. 2017, 13, 3031–3048. [Google Scholar] [CrossRef] [PubMed]
  33. Heo, L.; Feig, M. PREFMD: A web server for protein structure refinement via molecular dynamics simulations. Bioinformatics 2018, 34, 1063–1065. [Google Scholar] [CrossRef] [Green Version]
  34. Rohl, C.A.; Strauss, C.E.M.; Misura, K.M.S.; Baker, D. Protein structure prediction using Rosetta. Methods Enzymol. 2003, 383, 66. [Google Scholar]
  35. Wang, D.; Geng, L.; Zhao, Y.J.; Yang, Y.; Huang, Y.; Zhang, Y.; Shen, H.B. Artificial intelligence-based multi-objective optimization protocol for protein structure refinement. Bioinformatics 2020, 36, 437–448. [Google Scholar] [CrossRef] [PubMed]
  36. Moore, J.; Chapman, R.; Dozier, G. ACM Press the 38th annual. In Proceedings of the 38th Annual on Southeast Regional Conference, ACM-SE 38, Multiobjective Particle Swarm Optimization, Clemson, SC, USA, 7–8 April 2000; p. 56. [Google Scholar]
  37. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  38. Coello, C.A.C.; Pulido, G.T.; Lechuga, M.S. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  39. Hui, L.; Qingfu, Z. Multiobjective Optimization Problems with Complicated Pareto Sets, MOEA/D and NSGA-II. IEEE Trans. Evol. Comput. 2009, 13, 284–302. [Google Scholar] [CrossRef]
  40. Trivedi, A.; Srinivasan, D.; Sanyal, K.; Ghosh, A. A Survey of Multiobjective Evolutionary Algorithms based on Decomposition. IEEE Trans. Evol. Comput. 2016, 21, 440–462. [Google Scholar] [CrossRef]
  41. Zhang, J.; Liang, Y.; Zhang, Y. Atomic-level protein structure refinement using fragment-guided molecular dynamics conformation sampling. Structure 2011, 19, 1784–1795. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Cozzetto, D.; Kryshtafovych, A.; Fidelis, K.; Moult, J.; Tramontano, A. Evaluation of template-based models in CASP8 with standard measures. Proteins Struct. Funct. Bioinform. 2009, 77 (Suppl. 9), 18–28. [Google Scholar] [CrossRef] [Green Version]
  43. Zhang, Y.; Skolnick, J. TM-align: A protein structure alignment algorithm based on the TM-score. Nucleic Acids Res. 2005, 33, 2302–2309. [Google Scholar] [CrossRef]
  44. Qingfu, Z.; Hui, L. MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  45. Cheng, R.; Jin, Y.; Olhofer, M.; Sendhoff, B. A Reference Vector Guided Evolutionary Algorithm for Many-Objective Optimization. IEEE Trans. Evol. Comput. 2016, 20, 773–791. [Google Scholar] [CrossRef] [Green Version]
  46. Mohammadi, A.; Omidvar, M.N.; Li, X.; Deb, K. Sensitivity analysis of Penalty-based Boundary Intersection on aggregation-based EMO algorithms. In Proceedings of the 2015 IEEE Congress on Evolutionary Computation (CEC), Sendai, Japan, 25–28 May 2015. [Google Scholar]
  47. Yang, S.; Jiang, S.; Jiang, Y. Improving the Multiobjective Evolutionary Algorithm Based on Decomposition with New Penalty Schemes. Soft Comput. 2016, 21, 4677–4691. [Google Scholar] [CrossRef] [Green Version]
  48. Das, I.; Dennis, J.E. Normal-Boundary Intersection: A New Method for Generating the Pareto Surface in Nonlinear Multicriteria Optimization Problems. Siam J. Opt. 1996, 8, 631–657. [Google Scholar] [CrossRef] [Green Version]
  49. Branke, J.; Deb, K.; Dierolf, H.; Osswald, M. Finding knees in multi-objective optimization. In International Conference on Parallel Problem Solving from Nature; Springer: Berlin/Heidelberg, Germany, 2004. [Google Scholar]
  50. Parsons, J.; Holmes, J.B.; Rojas, J.M.; Tsai, J.; Strauss, C.E. Practical conversion from torsion space to Cartesian space for in silico protein synthesis. J. Comput. Chem. 2005, 26, 1063–1068. [Google Scholar] [CrossRef]
  51. AlQuraishi, M. Parallelized Natural Extension Reference Frame: Parallelized Conversion from Internal to Cartesian Coordinates. J. Comput. Chem. 2019, 40, 885–892. [Google Scholar] [CrossRef] [PubMed]
  52. Zhang, M.; Kavraki, L.E. A New Method for Fast and Accurate Derivation of Molecular Conformations. J. Chem. Inform. Model. 2002, 42, 64–70. [Google Scholar] [CrossRef] [Green Version]
  53. Tripathi, P.K.; Bandyopadhyay, S.; Pal, S.K. Multi-Objective Particle Swarm Optimization with time variant inertia and acceleration coefficients. Inform. Sci. 2007, 177, 5033–5049. [Google Scholar] [CrossRef] [Green Version]
  54. Zapotecas Martínez, S.; Coello Coello, C.A. A multi-objective particle swarm optimizer based on decomposition. In Proceedings of the 13th Annual Conference on Genetic and Evolutionary Computation, Dublin, Ireland, 12–16 July 2011; p. 69. [Google Scholar]
  55. Miettinen, K. Nonlinear Multiobjective Optimization; Springer: Berlin/Heidelberg, Germany, 1998. [Google Scholar]
  56. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Book Particle Swarm Optimization; BoD—Books on Demand GmbH: Norderstedt, Germany, 2002. [Google Scholar]
  57. Parsopoulos, K.E.; Vrahatis, M.N. Particle swarm optimization method in multiobjective problems. In Proceedings of the 2002 ACM Symposium on Applied Computing, Madrid, Spain, 10–14 March 2002. [Google Scholar]
  58. Wallner, B.; Fang, H.; Elofsson, A. Automatic consensus-based fold recognition using Pcons, ProQ, and Pmodeller. Proteins Struct. Funct. Bioinform. 2003, 53, 534–541. [Google Scholar] [CrossRef]
  59. Kim, H.; Kihara, D. Detecting local residue environment similarity for recognizing near-native structure models. Proteins 2014, 82, 3255–3272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Hiranuma, N.; Park, H.; Baek, M.; Anishchenko, I.; Dauparas, J.; Baker, D. Improved protein structure refinement guided by deep learning based accuracy estimation. Nat. Commun. 2021, 12, 1340. [Google Scholar] [CrossRef] [PubMed]
  61. Uziela, K.; Menendez Hurtado, D.; Shu, N.; Wallner, B.; Elofsson, A. ProQ3D: Improved model quality assessments using deep learning. Bioinformatics 2017, 33, 1578–1580. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Effectiveness of AIR 2.0 on CASP13 measured by GDT-TS, TM-score, and RMSD. The comparison of the best model refined by AIR 2.0 and the initial model in terms of GDT-TS, TM-score, and RMSD are shown in (ac), respectively. The comparison of AIR 2.0 refined Model 1 and the initial model in terms of GDT-TS, TM-score, and RMSD are given in (df) respectively.
Figure 1. Effectiveness of AIR 2.0 on CASP13 measured by GDT-TS, TM-score, and RMSD. The comparison of the best model refined by AIR 2.0 and the initial model in terms of GDT-TS, TM-score, and RMSD are shown in (ac), respectively. The comparison of AIR 2.0 refined Model 1 and the initial model in terms of GDT-TS, TM-score, and RMSD are given in (df) respectively.
Ijms 22 04408 g001
Figure 2. Refinement improvement over the initial model in terms of (a) GDT-TS (b) TM-score (c) RMSD.
Figure 2. Refinement improvement over the initial model in terms of (a) GDT-TS (b) TM-score (c) RMSD.
Ijms 22 04408 g002
Figure 3. The GDT-TS comparison of (a) best models and (b) Model1s between AIR 1.0 and AIR 2.0 on the targets of CASP13.
Figure 3. The GDT-TS comparison of (a) best models and (b) Model1s between AIR 1.0 and AIR 2.0 on the targets of CASP13.
Ijms 22 04408 g003
Figure 4. Pareto fronts of some targets obtained by AIR 2.0 (red) and AIR 1.0 (blue). (a) R0976D2, (b) R0981D5, (c) R0999D3, and (d) R1001. AIR 2.0 finds more non-dominated solutions than AIR 1.0, and these solutions are distributed with a high diversity on the target (b). Moreover, for targets such as (a,c,d), the solutions obtained by AIR 2.0 completely dominate those obtained by AIR 1.0, indicating a more convergence toward the true Pareto front.
Figure 4. Pareto fronts of some targets obtained by AIR 2.0 (red) and AIR 1.0 (blue). (a) R0976D2, (b) R0981D5, (c) R0999D3, and (d) R1001. AIR 2.0 finds more non-dominated solutions than AIR 1.0, and these solutions are distributed with a high diversity on the target (b). Moreover, for targets such as (a,c,d), the solutions obtained by AIR 2.0 completely dominate those obtained by AIR 1.0, indicating a more convergence toward the true Pareto front.
Ijms 22 04408 g004
Figure 5. The overall performance on 44 refinement targets of AIR in CASP14 blind test measured by GDT-TS.
Figure 5. The overall performance on 44 refinement targets of AIR in CASP14 blind test measured by GDT-TS.
Ijms 22 04408 g005
Figure 6. Candidate solutions obtained by AIR 2.0 (red) and AIR 1.0 (blue) on the target R0949. The Pareto front of R0949 is irregular and consists of at least two parts. The solutions of AIR 1.0 covers only one part, and the GDT-TS of its Model 1 marked with a green star is 62.98, which is a degradation to the initial model with a GDT-TS of 64.53. However, Model 1 of AIR 2.0 marked with a magenta star yields a GDT-TS of 65.12 on the other part of the Pareto front, demonstrating improvement over the initial model.
Figure 6. Candidate solutions obtained by AIR 2.0 (red) and AIR 1.0 (blue) on the target R0949. The Pareto front of R0949 is irregular and consists of at least two parts. The solutions of AIR 1.0 covers only one part, and the GDT-TS of its Model 1 marked with a green star is 62.98, which is a degradation to the initial model with a GDT-TS of 64.53. However, Model 1 of AIR 2.0 marked with a magenta star yields a GDT-TS of 65.12 on the other part of the Pareto front, demonstrating improvement over the initial model.
Ijms 22 04408 g006
Figure 7. Overall refinement protocol of AIR 2.0. The protocol consists of swarm initialization, multi-objective optimization with decomposition strategy, and solution ranking. In the second step, the different colored circles and stars denote particles and non-dominated solutions respectively.
Figure 7. Overall refinement protocol of AIR 2.0. The protocol consists of swarm initialization, multi-objective optimization with decomposition strategy, and solution ranking. In the second step, the different colored circles and stars denote particles and non-dominated solutions respectively.
Ijms 22 04408 g007
Figure 8. Illustration of PBI approach with a weight vector λ i = ( 0.33 ,   0.33 , 0.33 ) T . F(C) is the corresponding point of the candidate solution in the energy space. z * is the ideal objective vector. d 1 is the distance between the ideal objective vector z * and the solution F ( C ) , d2 is the direction error between λi and F(C). The Pareto front is represented by an orange plane. The PBI approach tries to minimize both d1 and d2. Obviously, the intersection of the weight vector and the Pareto front marked by a black point is the optimal solution of the subproblem defined by PBI with λi. It should be noted that the Pareto front is always irregular and discontinuous in practice. Here, we use a simple plane to represent it just for clarity.
Figure 8. Illustration of PBI approach with a weight vector λ i = ( 0.33 ,   0.33 , 0.33 ) T . F(C) is the corresponding point of the candidate solution in the energy space. z * is the ideal objective vector. d 1 is the distance between the ideal objective vector z * and the solution F ( C ) , d2 is the direction error between λi and F(C). The Pareto front is represented by an orange plane. The PBI approach tries to minimize both d1 and d2. Obviously, the intersection of the weight vector and the Pareto front marked by a black point is the optimal solution of the subproblem defined by PBI with λi. It should be noted that the Pareto front is always irregular and discontinuous in practice. Here, we use a simple plane to represent it just for clarity.
Ijms 22 04408 g008
Table 1. Overall performance of AIR 1.0 and AIR 2.0 in terms of GDT-TS on 29 refinement targets from CASP13.
Table 1. Overall performance of AIR 1.0 and AIR 2.0 in terms of GDT-TS on 29 refinement targets from CASP13.
MethodBest Model (GDT-TS)Model 1 (GDT-TS)
AIR 1.0 + 1.07 + 0.16
AIR 2.0 + 1.98 + 1.22
Table 2. GDT_TS comparison of AIR 2.0 and other refinement methods on CASP13 refinement targets. The results of BAKER, FEIGLAB, and Zhang come from the CASP official website. The best model among the four methods is bolded on each target.
Table 2. GDT_TS comparison of AIR 2.0 and other refinement methods on CASP13 refinement targets. The results of BAKER, FEIGLAB, and Zhang come from the CASP official website. The best model among the four methods is bolded on each target.
TargetInitial ModelAIR 2.0BAKERFEIGLABZhang
R094964.5365.1256.0162.9864.53
R095760.9764.0860.3261.4561.61
R0968s166.7469.7178.8172.2569.07
R0974s184.7885.9699.6497.1084.06
R0976D283.0684.2789.1180.6483.87
R097970.6574.1860.6070.3870.38
R0986s180.1683.1590.7693.2177.99
R0989D150.7551.1244.2250.75N/A
R0999D375.1476.9476.1176.9474.31
R1002D288.1488.5689.4179.2488.14
R1004D278.5777.6081.4993.5179.22
R101681.0682.1178.2281.6880.45
Table 3. Comparison of GDT-TS gains on different value of T .
Table 3. Comparison of GDT-TS gains on different value of T .
TargetLength T = 3 T = 8 T = 15 T = 30
R0974s169 0.271.180.811.24
R1004D277 1.05 0.97 0.98 0.95
R0968s11182.582.973.182.94
R0981D5127 0.790.590.440.20
R09591893.303.833.973.74
R0981D32030.490.1300.13
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhou, C.-P.; Wang, D.; Pan, X.; Shen, H.-B. Protein Structure Refinement Using Multi-Objective Particle Swarm Optimization with Decomposition Strategy. Int. J. Mol. Sci. 2021, 22, 4408. https://doi.org/10.3390/ijms22094408

AMA Style

Zhou C-P, Wang D, Pan X, Shen H-B. Protein Structure Refinement Using Multi-Objective Particle Swarm Optimization with Decomposition Strategy. International Journal of Molecular Sciences. 2021; 22(9):4408. https://doi.org/10.3390/ijms22094408

Chicago/Turabian Style

Zhou, Cheng-Peng, Di Wang, Xiaoyong Pan, and Hong-Bin Shen. 2021. "Protein Structure Refinement Using Multi-Objective Particle Swarm Optimization with Decomposition Strategy" International Journal of Molecular Sciences 22, no. 9: 4408. https://doi.org/10.3390/ijms22094408

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop