Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The Evolution of Your Success Lies at the Centre of Your Co-Authorship Network

Abstract

Collaboration among scholars and institutions is progressively becoming essential to the success of research grant procurement and to allow the emergence and evolution of scientific disciplines. Our work focuses on analysing if the volume of collaborations of one author together with the relevance of his collaborators is somewhat related to his research performance over time. In order to prove this relation we collected the temporal distributions of scholars’ publications and citations from the Google Scholar platform and the co-authorship network (of Computer Scientists) underlying the well-known DBLP bibliographic database. By the application of time series clustering, social network analysis and non-parametric statistics, we observe that scholars with similar publications (citations) patterns also tend to have a similar centrality in the co-authorship network. To our knowledge, this is the first work that considers success evolution with respect to co-authorship.

Introduction

Success, according to the Merriam Webster dictionary, is “getting or achieving wealth, respect, or fame”. But, how can we measure wealth, respect or fame and, ultimately, success? These concepts are subjective and, as Barabási claimed, success is a collective phenomenon in the sense that a person (or other entity) is successful because others around him believe that he is (http://www.northeastern.edu/news/2013/06/scienceofsuccess/). Despite this, some objective measures have been proposed for quantifying success, measures that depend on the domain (context) in which success is assessed. For instance, if we think about success in the marketing domain, a campaign will be successful if it gets to increase the profit. On the contrary, in social media, we could quantify the success of a video in YouTube by means of the number of views, the success of a tweet by the number of retweets received or the success of a Twitter user by the number of users who follow him (his followers).

In this paper we study academic researchers success: academic success can be tracked through the monitoring of the publication record of the academics in conferences and journals over the years. Many tools exist to gather data about published articles and citations and we will rely on these for our study. More specifically in this research we will focus on (i) the dissemination of results and (ii) the attention that these results receive in the research community as objective signs of scholars’ success. That is, we will focus on the publications (be conference papers, journals, books or patents) that authors get and the citations that these publications receive. We will also focus on scholars’ collaborations, and more specifically, on analysing if the volume of collaborations of one author, together with the relevance of his collaborators, is somewhat related to his research performance over time (temporal success). Therefore, the two main innovative factors of this work are the temporal evolution and the relationship of one author’s success to the success of his collaborators:

  • Temporal evolution: an important factor when assessing the success of an author should be the variability in the diffusion of their achievements over time. We study the temporal variation of an author’s publications and citations record: the number is not the only important measure, but the pace of publication activity and the impact of this activity are also very relevant.
  • Collaboration: the collaboration among researchers [1, 2] or their institutions [35] drives researchers success. As an outstanding example, the recent study of Bellotti [5] indicates that, in order to be successfully funded, the interaction over the years with different research groups counts more than working in a large university. Although this study is limited to research projects in the physics discipline funded by the Italian Ministry of University and Research, a look at the collaboration clusters winning research grant calls seems to confirm this informally too.

In order to prove the relation between the volume of publications/citations of one author over time and the collaboration network of co-authors of his papers, we collected the temporal distributions of scholars’ publications and citations from the Google Scholar platform (http://scholar.google.com/) and the co-authorship network (of Computer Scientists) underlying the well-known DBLP bibliographic database (http://www.informatik.uni-trier.de/∼ley/db/). Although Google Scholar contains the temporal distributions of all the publications and citations of each author (at least the ones from venues indexed by Google Scholar), the fact that not all the authors have a profile in this platform hampers us from having the complete view of the real collaboration network (henceforth, co-authorship network). On the contrary, the DBLP bibliographic database, despite only including Computer Science researches, provides a complete view of the co-authorship network, both in terms of nodes (scholars) and edges. However, this bibliographic database does not include citations data and therefore cannot be used for analysing the problem by itself alone. As a solution, we combine the aforementioned datasets: we apply time series clustering on the citations data of Google Scholar obtaining authors clusters; we then consider a collaboration network extracted from DBLP and apply social network analysis on it. We observe that the (median) centrality of those scholars with similar publications (citations) patterns (i.e. in the same cluster) is different from the (median) centralities in the rest of clusters. These findings open the door to potentially new publishing strategies as well as prediction of success for young scholars, potential recruits as well as journal editors.

The main novelty of our approach is to consider the whole scholars’ timeline as sign of their research activity. Although both citation and publication counts have been widely used as indicators of scholars’ impact [68], most of the existing metrics only consider one time window which may be difficult to select as it depends on several cross-cutting factors [9, 10]. Given the importance of time of publications/citations in determining success, the complete timeline of the scholars may give a more complete view. Moreover, our work establishes scholars’ collaboration according to their relative importance within the co-authorship network (centrality) as the most formal sign of academic teamwork. In order to establish the dependence among the longitudinal data (time series of citations and publications) by obtaining some derivable variable from the series, we propose retaining all the information of the temporal research activity and exploring temporal patterns for scholars by clustering their timelines. Then, after studying the degree, closeness, betweenness and eigenvector centrality measures considering the collaboration network for the clustered authors and observing that they are far from being normally distributed, we use non-parametric statistics (Kruskal-Wallis statistical test [11]) in order to check whether there are differences, in terms of authors’ centrality, among the different clusters. That is, whether the median centrality of the authors classified in the same cluster is different from the median centralities in the rest of clusters.

The main findings of our study are:

  • The volume of publications/citations over time of one Computer Scientist who started his career between 1979 and 2009 is related to the volume and relevance of his collaborators. That is, to his centrality in the co-authorship network.
  • This relation holds for most of the period considered, which means that collaboration affects (and is affected by) research performance over time. What is likely is that, the more collaborations one author has, the higher his number of publications will be and the more attention they will receive.

Materials and Methods

Datasets collection

We now describe the two datasets we have worked with and the reasoning behind this choice.

  • The Google Scholar platform allows academics to create profile pages that contain, in addition to their affiliation and areas of interest, their history of publications and citation counts over time. A profile page in this platform allows other academics to see, at a glance, the author’s publications without having to search in the traditional Google Scholar page. Also, and more importantly, instead of only showing the total number of citations of a paper (author), it displays the citations distribution of the paper (author) over time. Google Scholar also allows researchers to link to collaborators, however this function is not very popular and therefore it is almost impossible to gather collaboration network data from Scholar.
  • To complement Google Scholar and be able to obtain collaboration network information, we used the DBLP Computer Science Bibliography, a tool developed by researches from the University of Trier: with this, we trace co-authorship in the work of Computer Science researchers. Although considering only Computer Science researchers limits the analysis, this gives us a complete network of co-authors to work with.

We now describe how we obtained the data.

Google Scholar

Every author in Google Scholar has his own identifier, but there is no way to know all of the identifiers of Google Scholar authors. So, one indirect way for getting the given identifiers (or at least most of them) consists of crawling Google Scholar through the connections among the co-authors that authors have previously included in their profiles. But, as aforementioned, the number of authors that have explicitly indicated their co-authors is very small. However, many of them do have indicated their areas of research. As Google Scholar allows to query authors interested in a given area, if we knew all the areas of interest in Google Scholar, we could get all authors’ identifiers (at least the ones that have indicated their areas of interest). Again, as authors can freely indicate their areas of expertise (there is not a predefined group of them) there is not direct way to know all the areas of interest in Google Scholar. We have devised an algorithm that, starting with an author’s profile (acting as seed), retrieves his areas of interest and all the authors that have indicated the same areas as him. It then randomly selects one of these authors and repeats this procedure, retrieving the areas and the authors that have not been crawled yet. The algorithm finishes when there are not more authors to select.

Our dataset contains information about 192,930 authors with profile in Google Scholar and, at least, one publication indexed by Google Scholar and 9,030,060 papers. Table 1 summarises (i) authors’ and (ii) papers’ attributes contained in this dataset (previously retrieved from the Google Scholar platform).

DBLP

The DBLP Computer Science Bibliography is a tool developed by researches from the University of Trier to trace the work of Computer Science researchers. The whole DBLP content is available, in form of XML file, in http://dblp.uni-trier.de/xml, whose structure is described in http://dblp.uni-trier.de/xml/dblp.dtd. An updated version of the XML file is released daily. As of 2013, the dataset includes 1,227,149 authors and 9,822,354 co-authorship relations.

The information in this dataset is presented in a publication-centred perspective. Different publications are considered, following a Bibtex style, as: article, inproceedings, proceedings, book, incollection, phthesis, mastherthesis or www. Data associated to each publication are shown in Table 2. The relevant information for our purposes is authors’ names and publications, which will allow us to obtain the co-authorship network. That is, the network in which nodes are the computer science authors and a link between them exists if they have co-authored, at least, one publication.

The aforementioned datasets (Google Scholar and DBLP) have different levels of coverage: while Google Scholar contains information about researchers in whatever area of expertise and who have created a profile in this platform, DBLP contains information about all the researchers in the Computer Science domain (or at least about those that have published, at least once, in a conference/journal indexed by DBLP). The two platforms associate users to different identifiers which means that we have to match authors in both platforms. To this aim, we applied a conservative solution: the lexicographic comparison of authors’ names in both datasets, obtaining that the number of authors with profile in both datasets is 57416, which represents around 30% of our entire Google Scholar dataset.

Analysis Techniques

As mentioned, one of the main contributions of this work is the ability to track the evolution of authors’ success through the use of the temporal data contained in our datasets. We analyze the data longitudinally without resorting to aggregates or univariate summaries (e.g. averages or time slopes). Our approach can be briefly described as follows: we first apply hierarchical clustering to the Google Scholar longitudinal data containing citations and publications, operation that divides our authors into groups with similar patterns of publications (citations). We then consider the DLBP collaboration network and apply social network analysis to determine author centrality. Finally we study the relation between the author centralities and the clusters. In the next sections we detail the various steps we just described.

Time Series Analysis

We start by exploring the longitudinal dataset to discover groups of authors that share similar publication/citation dynamics as a mean to identify different patterns of success evolution in the pool of authors. For that, among unsupervised techniques for exploratory data analysis, clustering is a strong candidate for finding strongly related data points. An ample variety of algorithms have been proposed for clustering purposes (see [12] for a review). Given the scarce knowledge about temporal patterns in scholars’ activity to date and considering its conceptual simplicity and good scaling with a large number of points, a hierarchical clustering algorithm (The Unweighted Pair Group Method with Arithmetic Mean (UPGMA) [13]) seems appropriate as it constructs a rooted tree (dendrogram) for inspection. This dendrogram reflects the structure present in a pairwise similarity matrix (or a dissimilarity matrix) by accomplishing iterative steps as follows:

  1. Place each data point into its own singleton group.
  2. Merge the two closest groups.
  3. Update distances D between the new cluster (group) and each of the old clusters (see below).
  4. Repeat 2 and 3 until all the data are merged into a single cluster.

Given a distance measure d between points UPGMA obtains the intergroup similarity between the clusters C and H as: (1) where NC(NH) is the size of the cluster C(H) and di,j is the distance between the data points i and j.

For the case of UPGMA over time series, the pairwise similarity matrix is composed of similarity measures between every two time series (authors’ time series of publications or citations as corresponds). Our methodology constructs this matrix by applying the well-known Dynamic Time Warping (DTW) method [14], which finds an optimal alignment between two time-dependent sequences S1 and S2 by warping the time dimension in S1 that minimises the difference between the two series so that time series are not need to be of equal length. Although DTW was initially applied to word’s recognition continuous human speech [15, 16], its use has been extended to other domains. In our experiment, we alleviate the problem of alignment by classifying authors in advance by their first publications’s year, so that only time series of equal length are considered.

DTW works as follows. Let us suppose that we have two time series S1 = {s11, s12, …, s1n} and S2 = {s21, s22, …, s2m} that represent the temporal distribution of publications (citations) of author 1 and author 2 respectively. If i = 1, 2, …n and j = 1, 2, …, m are indices into S1 and S2 and W = w1, w2, …, wK, is an optimal mapping between the two sequences given as, (2) we can compute the dissimilarity or distance between the two sequences as, (3) where δ(wk) is a non-negative function to compute dissimilarity between individual elements of S1 and S2.

As most clustering algorithms, the key issue in UPGMA is fixing the number of resulting clusters. Being UPGMA a hierarchical method, this turns into the problem of identifying the relevant branches of the cluster tree (dendrogram pruning). In order to not have to decide for each individual branch, the most widely used technique is based on fixing the height of the dendrogram so that each contiguous branch below that height is considered a separate cluster. Unfortunately, selecting this height is not a trivial task and, alternatively, Dynamic Tree Cut (DTC) [17] proposes to prune the dendrogram by taking its shape into consideration. DTC iteratively applies decomposition and combination of clusters until their number becomes stable. After obtaining a few large clusters by the fixed height branch cut method, the joining heights of each cluster are analysed for a sub-cluster structure. Clusters with this sub-cluster structure are recursively split and, with the aim of avoiding over-splitting, very small clusters are joined to their neighbouring major clusters.

Social Network Analysis

Once we have discovered the structure of the temporal dataset, the next step is to establish if this structure is related to scholars’ centrality in the co-authorship network. For that, we take advantage of the DBLP dataset to obtain the current network of co-authorship. We represent this network as an undirected graph G = (V, E), where V denotes the set of scholars (authors) in the dataset, and an edge eE between two scholars u, v ∊ V exists if they have co-authored, at least, one publication. We will then be able to determine if the position of the author in the given network has relation with his citations (publications) behaviour in Google Scholar. That is, if the fact that two authors are included in the same cluster implies that they have similar centrality (we have tried various measures of centrality) in the network. This will indicate if researchers’ collaboration activity is dependent on the temporal distribution of the publications or the citations that they receive. With this aim, we applied social network analysis over the co-authorship network extracted from DBLP.

The next step consists in determing the centrality of each node within the network. Several measures of centrality have been proposed to date. The most fundamental and popular definitions of centrality were proposed by Freeman [18]. By this definition, the centrality is measured based on degree, closeness and betweenness. These measures, together with the eigenvector centrality, are detailed below.

  • Degree centrality: The degree of a node ni in a graph is the number of direct connections that a node has. It is the simplest and easiest way of measuring its centrality, being a local measure of a node’s importance. The degree centrality of a node reflects the popularity and relational activity of the node. (4) where N is the number of nodes in the network and a(ni, nk) is a distance function. a(ni, nk) = 1, if and only if node ni and node nk are connected. a(ni, nk) = 0 otherwise.
  • Betweenness centrality: The betweenness centrality of a node ni is defined to be the fraction of shortest paths between pairs of nodes in a network that pass through ni. It represents the node’s capability to influence or control interaction between nodes that it links. (5) where gnjnk (ni) is the number of shortest paths connecting nj and nk passing through ni and gnjnk is the total number of shortest paths connecting them.
  • Closeness centrality: The closeness centrality of a node measures how many steps are required to reach a given node from every other node. It indicates the node’s availability, safety and security depending on the application context considered. (6) where N is the number of nodes in the network and d(ni, nk) is the number of hops in the shortest path between ni and nj.
  • Eigenvector centrality: The eigenvector centrality is a measure of the importance of a node in a network and it is based on the idea that a node is more central if it is in relation with nodes that are themselves central. (7) where λ is a constant, N is the number of nodes in the network and a(ni, nk) is a distance function. a(ni, nk) = 1, if and only if node ni and node nk are connected. a(ni, nk) = 0 otherwise.

The next section shows the results of our analysis.

Results

Our methodology for analysing the data obtained from both Google Scholar and DBLP processes each dataset separately. Then, the results are analyzed through statistical tests to determine if there exists any relation between them. In the case of Google Scholar, we also process separately publications and citations, as they are different indicators of authors’ success: whereas a publication is reviewed by a group of previously selected experts in the domain (reviewers), any author, in or out of the domain, can cite a previously published publication. On the other hand, authors have started to publish (receive citations) in different years and their publications and citations are conditioned by the state of the research environment in every moment of their trajectory. The number of existing conferences and journals or the accessibility to bibliographic resources are only some of the factors that condition authors’ publications and citations in a given period. To avoid the impact of this aspect on the results, we decided to split authors in groups depending on the year in which they started to publish (in the case of publications) or to receive the first citation (in the case of citations). In the case of DBLP, as the co-authors network is the result of collaborations between authors through the years, we work with the whole network.

With respect to the distribution of authors by year of their first publication (citation), Fig. 1A represents the number of authors classified by (i) the year in which they achieved their first publication and (ii) the year in which any of their publications received its first citation. Both the Computer Science discipline and the Google Scholar platform are relatively recent in comparison to other scientific disciplines and other scientific databases, which could explain why the curves of the number of authors per year are increasing.

thumbnail
Fig 1. Time series clustering in Google Scholar and Centrality measures in DBLP.

(a) Number of authors attending to the year of their first publication (citation). (b) Number of clusters of authors per year. (c) Number of authors per cluster per year (attending to the temporal evolution of their publications). (d) Number of authors per cluster per year (attending to the temporal evolution of their citations). (e) Degree centrality in the co-authorship network (for authors simultaneously in both datasets and for authors in the entire DBLP). (f) Betweenness centrality in the co-authorship network (for authors simultaneously in both datasets and for authors in the entire DBLP). (g) Closeness centrality in the co-authorship network (for authors simultaneously in both datasets and for authors in the entire DBLP). (h) Eigenvector centrality in the co-authorship network (for authors simultaneously in both datasets and for authors in the entire DBLP).

https://doi.org/10.1371/journal.pone.0114302.g001

Grouping authors

As aforementioned, we applied hierarchical clustering over the time series of authors’ publications (citations) in order to extract groups of authors with similar publication (citation) patterns. Fig. 1 contains information about the resulting clusters: number of resulting clusters (Fig. 1B) and number of authors per cluster in the case of publications (Fig. 1C) and citations (Fig. 1D).

Focusing on Fig. 1B, we see that the number of resulting clusters increases with the year of the first publication (citation) and, therefore, with the number of time series (authors) to be clustered, at least until the late 2000s. That is, the more authors to cluster, the more different patterns emerge. Although this tendency is notable both in publications and citations, the increase is higher in the case of publications, highlighting the variability of the publication records. With respect to the number of authors (time series) per cluster, we appreciate that the resulting clusters have, in average and independently of the year, equal number of elements (authors). On the contrary, the curve of the maximum number of elements in a cluster is increasing with the number of authors in a cluster until mid-2000s. When there are more authors who started to publish (receive citations) in the same year, (i) the diversity in terms of publication/citation patterns is higher and (ii), although in average the number of authors with similar patterns (classified in the same cluster) keeps constant with respect previous years, there are some big groups of authors (clusters) that share a similar pattern of publications (citations).

However, and as aforementioned, there are certain periods in mid-2000s in which, although (i) the number of authors who started their activity around these years and (ii) the diversity in terms of citation/publication patterns are higher than in previous years, there are less authors that share similar patterns (less authors per cluster). A feasible explanation for this could be that, as these authors are relatively new to the research area, they are in a “transitory” period of their careers, which could cause the high diversity of behaviour among them. Finally, Fig. 1 also shows that recent scholars, those who start to publish/being cited between 2010 to 2012 and therefore have had less than 3 years to develop their research activity, are, in general, authors with very few publications/citations which makes that almost all of them present the same publication/citation pattern.

Obtaining authors’ centrality

The CCDF (Complementary Cumulative Distribution Function) curves corresponding to the centrality measures previously described are included in Fig. 1. For each measure, we provide two curves: the one that represents the distribution of the given centrality for the authors in the whole DBLP dataset and the other that limits the curve to only those authors that simultaneously appear in DBLP and in Google Scholar dataset (those subject of our study).

Specifically, Fig. 1E shows the distribution of degree centrality among authors, that is, the number of authors with whom they have collaborated at least once. Although the tendency of the curve is similar in the case of all computer science authors or only computer science authors with profile in Google Scholar, the degree is slightly higher in the second case. Looking at the CCDF for the whole dataset, we see that around the 40% of the authors have a degree centrality value higher than 5, whereas when only authors in our Google Scholar dataset are considered this percentage increases to the 70%. This means that authors who own a profile in Google Scholar are more prone to co-author papers with different authors than the rest of computer science authors. In the case of eigenvector centrality (Fig. 1H) most of the nodes (authors) in the network have an eigenvector centrality equal to zero. That is, around the 11% of the authors in the whole dataset have a positive eigenvector centrality, whereas this percentage increases to the 27% when only authors in Google Scholar are taken into account. Talking in terms of research collaborations, this means that scholars are prone to collaborate with both, central and non-central authors, which means that collaborations are quite distributed across the network instead of being topologically focused on only one or a few areas in the network.

In the case of betweenness centrality (Fig. 1F), around the 30% of authors that are simultaneously in both datasets (Google Scholar and DBLP) have no shortest path through them, whereas this percentage increases up to almost 60% when the entire DBLP dataset is taken into account. This suggests that authors who own a profile in Google Scholar are more prone to serve as links between scholars (co-authorise papers with well-connected authors) than the rest of computer science authors. With respect to authors with positive values of betweenness, there is no difference between the distribution of this variable among the total authors in computer science (in the whole DBLP dataset) with respect to the distribution when only Computer Scientists with a profile in Google Scholar are considered. Finally, in the case of closeness centrality (Fig. 1G) the distribution of closeness centrality of authors in Google Scholar is similar to the one of all computer science authors. Specifically, the total of computer science scholars have a value of closeness higher than 0.12 and only the 10% of them have a value of closeness equal to 1, being this percentage reduced until the 5% in the case of authors with profile in Google Scholar. Talking in terms of co-authorship relations, this means that only a reduced percentage of scholars are at a reachable distance, in terms of co-authorship links, of the rest of computer science authors. But, the immense majority of authors is far from having collaborated with the rest of nodes in the network.

Relating academic success and centrality: the Kruskal-Wallis test

The last step of our analysis deals with checking the existence of relation among the authors included in the same cluster and their centrality in the co-authorship network. To this aim, the common approach in the literature is the one-way analysis of variance [19] (one-way ANOVA), a statistical technique used to compare the means of two or more samples (being the samples, in our case, the authors classified into each cluster). Specifically, ANOVA tests the null hypothesis that samples in two or more groups are drawn from populations with the same mean values. In our scenario, this implies testing if the authors classified in the previously calculated clusters (response or dependent variable) are drawn from authors with the same mean values of centrality in the co-authorship network (factor or independent variable). But the reliability of one-way ANOVA results is conditioned, among others assumptions, by normality of the response variables, and the application of the Shapiro-Wilk statistical test [20] revealed that centrality distribution of authors within a cluster is far for being normally distributed. Under these circumstances, we opted for a non-parametric alternative to the one-way ANOVA, the Kruskal-Wallis test [11]. Contrary to one-way ANOVA, which worked with mean values, Kruskal-Wallis tests the null hypothesis that samples in two or more groups are drawn from populations with the same median values. Thus, we applied Kruskal-Wallis test to determine if the centralities of the authors with the same publication (citation) pattern had the same median value and if this value was different from the median value of authors with different publication (citation) patterns. That is, if authors with similar centrality in the network have similar publication (citation) patterns and these patterns are different from the ones from authors with different centrality values.

The Kruskal-Wallis statistical results considering degree, betweenness, closeness and eigenvector centrality measures for the different years are shown in Tables 3 and 4 for publications and citations respectively. Taking the standard α = 0.05 as the significance level for all the Kruskall-Wallis tests conducted, results revealed that the p-value is lower than the significance value, and therefore the null hypothesis can be rejected for the most relevant cases, those which includes authors whose starting publication date was between 1979 and 2009 (Table 3), for all the centrality with the exception of the closeness centrality during some years between 1979 and 1987. These results are highly conclusive since, for authors who started to publish between 2010 and 2012 (classified into two or more clusters attending to their publications rate), the length of their time series (2 or 3 points) is not enough for obtaining accurate values of similarity and, consequently, for a suitable (clustering) classification. A similar situation can be appreciated in authors that started to publish before 1978. In this period, there exists a huge diversity, both in terms of measures and years, in which the null hypothesis cannot be rejected. In any case, and due to the peculiarities of the Google Scholar dataset (that counts publications since 1974) and that Computer Science is a relatively recent discipline, the number of authors that started to publish before 1980 is quite limited with respect to the number of computer scientists that started their activity later and these results should not be considered relevant. Again, clustering for this period may not be considered reliable enough. With all of this, the conclusions so far are that (i) in around the 80% of the considered period authors’ centrality is highly related with the distribution of publications over time and (ii) the periods in which this relation is not determined correspond to periods in which there are very few authors or periods of temporal series of publications extremely short.

thumbnail
Table 3. Kruskal-Wallis statistics results for the publications case.

https://doi.org/10.1371/journal.pone.0114302.t003

thumbnail
Table 4. Kruskal-Wallis statistics results for the citations case.

https://doi.org/10.1371/journal.pone.0114302.t004

A similar tendency is appreciated in the case of citations (Table 4). Specifically, the null hypothesis can be rejected when considering authors that start to receive citations between 1988 and 2012, with exceptions during some years, again when closeness centrality is considered. However, the interval in which we cannot deduce anything about the relation between citation patterns and centrality is extended to authors that started to receive citations between 1974 and 1988. Finally, similar conclusions are drawn when considering different measures of centrality, where there are not differences between authors who started to receive citations in one year or another (except when considering authors that started to receive citations in 1984).

Comparisons with a null model

Kruskal-Wallis using a random graph as response.

In order to demonstrate the significance of the results achieved, we used the Kruskal-Wallis test on a shuffled DBLP network: instead of considering the real edges in the DBLP network, we modeled these edges by means of a classic random network model, the Erdös-Rényi random graph, G(n, p) [21]. This graph, G(n, p) is defined by two parameters: the number of nodes in the graph (n) and the edge probability for drawing an edge between two arbitrary nodes (p). Our random network was obtaining using n = 57416 (equal to the number of authors simultaneously in both datasets) and p = 1.4e–4 (in order to have a similar level of connection than the original DBLP network), although with different distribution (random). With this test we wanted to exclude the correlation of random collaboration network distributions with our clusters. Results achieved are shown in Tables 5 and 6. According to these tables, both for the publications and citations cases, the null hypothesis cannot be rejected in almost all the the years (in which authors started to publish/receive citations) and for all the different centrality measures considered, which confirms the significance of our results.

thumbnail
Table 5. Kruskal-Wallis statistics results for the publications case (using a random graph as a response).

https://doi.org/10.1371/journal.pone.0114302.t005

thumbnail
Table 6. Kruskal-Wallis statistics results for the citations case (using a random graph as a response).

https://doi.org/10.1371/journal.pone.0114302.t006

Correlation measures between authors’ centrality and their publications (citations) counts.

With the aim of testing the relation between authors’ centrality in the co-authors network and their total number of publications (citations), we calculated three well-known correlation indexes: Pearson, Spearman and Kendall. Fig. 2 shows the values of the different indexes taking into account different centrality measures (degree, betweenness, closeness and eigenvector centrality) and splitting authors according to the year of their first publication/citation. Results revealed that the correlation between centrality and number of publications (citations) is not significative (lower than 0.5 in almost cases), with the only exception of the authors who started their activity around 1990 when considering the correlation between their publications and degree (betweenness) centrality (Fig. 2A and Fig. 2B). This justifies the necessity of considering the whole scholars’ timeline as sign of their research activity.

thumbnail
Fig 2. Correlation measures between authors’ centrality and their publications (citations) count.

(a) Pearson correlation between authors’ centrality and their publications count. (b) Spearman correlation between authors’ centrality and their publications count. (c) Kendall correlation between authors’ centrality and their publications count. (d) Pearson correlation between authors’ centrality and their citations count. (e) Spearman correlation between authors’ centrality and their citations count. (f) Kendall correlation between authors’ centrality and their citations count.

https://doi.org/10.1371/journal.pone.0114302.g002

Other experiments without considering time. Finally, we run an experiment to demonstrate the importance of classifying authors according to the year of their first publications/citation when looking for their relation with centrality in the co-authorship network. Specifically, we computed the Pearson, Spearman and Kendall correlation indexes between the number of publications/citations and the degree centrality (the one with best results in Fig. 2) of all the authors in the dataset (independently of the length of their careers). Results in Table 7, with correlation values within the ranges of the ones in Fig. 2 confirm the importance of splitting authors by the year in which their careers started.

thumbnail
Table 7. Correlation measures between authors’ centrality and their publications (citations) counts without splitting authors per year of their first publication (citation).

https://doi.org/10.1371/journal.pone.0114302.t007

Discussion

The main focus of this work is on analysing the importance of collaborations in academic success over time, by considering success as the volume of publications and cites to these publications. That is, on analysing if the volume of collaborations of one author together with the relevance of his collaborators is related to his research performance over time. We made use of two different datasets: one obtained by crawling the Google Scholar platform and the other the well-known DBLP dataset. By the application of a two-steps methodology based on (i) clustering scholars’ timelines to explore their temporal patterns and (ii) the application of non-parametric statistics to establish the correlation among timeline and centrality, we confirmed our hypothesis that computer scientists’ centrality in the co-authorship network is related with the patterns that their publications (citations) followed along the years. Although this relation was found both in the case of publications and citations and independently of the centrality measure, we could not guarantee the existence of relation (i) for scholars that started to publish/receive citations recently (very short time series) neither (ii) for those that started to publish before 1979 in the case of publications (1988 in the case of citations). This relation holds for most of the period considered: this confirms our initial hypothesis that collaboration affects (and is affected by) research performance over time. What is likely is that, the more collaborations one author has, the higher his number of publications will be and the more attention they will receive. Moreover, the relevance of these collaborations seems to play a key role in the increase the number of publications and their visibility.

With respect to future work, our analysis could be extended in three different ways. Firstly, our mechanism for matching authors in the datasets could be improved by (i) taking into account measures of lexicographic distance between strings as the Levenshtein distance and (ii) applying techniques to detect ambiguous author names in scholarly data as the one proposed by Sun et al. [22, 23]. At the moment authors with the same name are considered the same person while by using the aforementioned techniques, we expect to increase the number of authors simultaneously in both datasets and, therefore, increasing the size of the sample, besides avoiding false positives in the matching-mechanism. Secondly, by considering other network measures, such as cross-clustering coefficient and cross-transitivity [24, 25] to conduct social network analysis to see whether more further information could be uncovered. Finally, once temporal series are classified in clusters, it could be interesting to characterise those series according to the different types of continuous dynamics that they exhibit: periodic, chaotic, and periodic with noisy [2628].

To conclude, the extension of our study to other disciplines besides Computer Science would give us a better understanding of the dynamics of research. Once discovered these dynamics, the next logical step should be the development of a prediction framework that allows scholars to improve their collaboration network in order to increase their temporal success.

Supporting Information

S1 Appendix. Anatomy of the Google Scholar dataset.

https://doi.org/10.1371/journal.pone.0114302.s001

(PDF)

Acknowledgments

This work was partially done while Sandra Servia-Rodríguez was visiting the Computer Laboratory of the University of Cambridge.

Author Contributions

Conceived and designed the experiments: SS AN CM AF RPD. Performed the experiments: SS. Analyzed the data: SS. Contributed reagents/materials/analysis tools: SS. Wrote the paper: SS AN CM AF RPD.

References

  1. 1. Newman M (2004) Coauthorship networks and patterns of scientific collaboration. Proceedings of the National Academy of Sciences 101: 5200–5205. pmid:14745042
  2. 2. Sun X, Kaur J, Milojevi S, Flammini A, Menczer F Social Dynamics of Science. Scientific Reports 3: DOI:10.1038/srep01069.
  3. 3. Deville P, Wang D, Sinatra R, Song C, Blondel VD, et al. (2014) Career on the Move: Geography, Stratification, and Scientific Impact. Scientific reports 4. pmid:24759743
  4. 4. Gargiulo F, Carletti T (2014) Driving forces of researchers mobility. Scientific reports 4. pmid:24810800
  5. 5. Bellotti E (2012) Getting funded. Multi-level network of physicists in Italy. Social Networks 34: 215–229.
  6. 6. Ding Y, Yan E, Frazho A, Caverlee J (2009) Pagerank for ranking authors in co-citation networks. Journal of the American Society for Information Science and Technology 60: 2229–2243.
  7. 7. Abbasi A, Altmann J, Hossain L (2011) Identifying the effects of co-authorship networks on the performance of scholars: A correlation and regression analysis of performance measures and social network analysis measures. Journal of Informetrics 5: 594–607.
  8. 8. Abbasi A, Chung KSK, Hossain L (2012) Egocentric analysis of co-authorship network structure, position and performance. Information Processing & Management 48: 671–679.
  9. 9. Wang J (2013) Citation time window choice for research impact evaluation. Scientometrics 94: 851–872.
  10. 10. Sarigöl E, Pfitzner R, Scholtes I, Garas A, Schweitzer F (2014) Predicting scientific success based on coauthorship networks. arXiv preprint arXiv:14027268.
  11. 11. Kruskal WH, Wallis WA (1952) Use of ranks in one-criterion variance analysis. Journal of the American statistical Association 47: 583–621.
  12. 12. Jain A, Murty N, Flynn P (1999) Data clustering: a review. ACM Computing Surveys (CSUR) 31: 264–323.
  13. 13. Jain A, Dubes R (1988) Algorithms for clustering data. Prentice-Hall, Inc.
  14. 14. Berndt D, Clifford J (1994) Using dynamic time warping to find patterns in time series. In: KDD workshop. Seattle, WA, volume 10, pp. 359–370.
  15. 15. Sakoe H, Chiba S (1978) Dynamic programming algorithm optimization for spoken word recognition. Acoustics, Speech and Signal Processing, IEEE Transactions on 26: 43–49.
  16. 16. Sakoe H, Chiba S (1971) A dynamic programming approach to continuous speech recognition. In: Proceedings of the seventh international congress on acoustics. volume 3, pp. 65–69.
  17. 17. Langfelder P, Zhang B, Horvath S (2008) Defining clusters from a hierarchical cluster tree: the Dynamic Tree Cut package for R. Bioinformatics 24: 719–720. pmid:18024473
  18. 18. Freeman L (1979) Centrality in social networks conceptual clarification. Social Networks 1: 215–239.
  19. 19. Miller RG Jr (1997) Beyond ANOVA: basics of applied statistics. CRC Press.
  20. 20. Shapiro SS, Wilk MB (1965) An analysis of variance test for normality (complete samples). Biometrika: 591–611.
  21. 21. Bollobás B (1998) Random graphs. Springer.
  22. 22. Sun X, Kaur J, Possamai L, Menczer F (2011) Detecting ambiguous author names in crowdsourced scholarly data. In: 2011 IEEE Third international conference on Privacy, security, risk and trust (PASSAT) and 2011 IEEE Third international conference on Social Computing (socialCom). IEEE, pp. 568–571.
  23. 23. Sun X, Kaur J, Possamai L, Menczer F (2013) Ambiguous author query detection using crowdsourced digital library annotations. Information Processing & Management 49: 454–464.
  24. 24. Gao ZK, Zhang XW, Jin ND, Marwan N, Kurths J (2013) Multivariate recurrence network analysis for characterizing horizontal oil-water two-phase flow. Physical Review E 88: 032910. pmid:24125328
  25. 25. Gao ZK, Zhang XW, Jin ND, Donner RV, Marwan N, et al. (2013) Recurrence networks from multivariate signals for uncovering dynamic transitions of horizontal oil-water stratified flows. EPL (Europhysics Letters) 103: 50004. pmid:24125328
  26. 26. Xu X, Zhang J, Small M (2008) Superfamily phenomena and motifs of networks induced from time series. Proceedings of the National Academy of Sciences 105: 19601–19605. pmid:19064916
  27. 27. Zhang J, Small M (2006) Complex network from pseudoperiodic time series: Topology versus dynamics. Physical Review Letters 96: 238701. pmid:16803415
  28. 28. Gao ZK, Jin ND (2012) A directed weighted complex network for characterizing chaotic dynamics from time series. Nonlinear Analysis: Real World Applications 13: 947–952.