Keywords
Journal rankings, ANVUR, Arts, Humanities and Social Sciences, Journal ratings, Article Quality
This article is included in the Research on Research, Policy & Culture gateway.
This article is included in the Proceedings of the 2015 ORCID-Casrai Joint Conference collection.
Journal rankings, ANVUR, Arts, Humanities and Social Sciences, Journal ratings, Article Quality
There is a large degree of agreement on the notion that research assessment in humanities and social sciences (HSS) is made more complex by a variety of factors. First, in these fields the structure of academic publication is largely different, with a large weight assigned to books and monographs, and to production in national language (Finkenstaedt, 1990). Consequently, the bibliometric approach is considered to be of limited usefulness (Nederhof et al., 1989), not only because journals are a small fraction of total production and indexed journals are a tiny fraction of the population of journals in HSS, but also because the meaning of citations is different (Frost, 1979). Third, and even more challenging, there is evidence that the number of research quality criteria is larger in HSS than in other fields, and also that there is less agreement on these criteria (Hemlin, 1996; Hemlin & Gustafsson, 1996; Hug et al., 2013; Hug et al., 2014; Ochsner et al., 2012; Ochsner et al., 2013).
Faced with these challenges, the state of the art of the assessment of research in HSS at international level has followed several directions. On the one hand, it is agreed that peer review is still the most important evaluation methodology, so large efforts are made in making it more sophisticated, methodologically controlled, based on sound principles of evaluation methodology in social sciences, and free from unwanted biases, distortions and unexpected side effects. Under this agenda, issues such as the notion of originality, unorthodox science, or interdisciplinarity are under examination (Guetzkow et al., 2004; Hammarfelt, 2011). On the other hand, there are many efforts to classify and evaluate non-indexed journals (mainly in national languages), as one of the main vehicles for academic communication. An additional line of work refers to the classification of books and publishers.
This paper reports on a large experiment in the classification of journals in HSS carried out in Italy in the 2012–2014 period for the National Scientific Habilitation (Abilitazione Scientifica Nazionale; ASN). The exercise was based upon a mandatory provision in the law to rate all journals, in order to calculate the overall academic production of all candidates to the national procedure to become associate professor or full professor. This exercise asked the National Agency for the Evaluation of Universities and Research Institutes (ANVUR) to evaluate all journals in which at least one Italian scholar published at least one paper in the 2002–2012 period, for a total of more than 60,000 titles.
While the rating of journals has been followed in several national contexts, it is only in the Italian exercise that there is the opportunity to carry out a controlled experiment in order to test the robustness of journal classification. In fact, we have two independent evaluations carried out on the same set of journals. On the one hand, a panel of experts classified all journals as academic and non-academic (i.e. popular, professional, technical, cultural and political etc.), and rated the subset of academic journals in A-rated and non A-rated. The rating exercise was done on the basis of the reputation, esteem, diffusion and impact of journals, that is, on a qualitative, expert-based, reputational basis. On the other hand, we also have the rating of individual articles published in those journals, which have been done by a large number of individual referees (not panels) and summarized with a consensus agreement approach by expert panels, who however acted independently from the other panels, and without exchange of information. This peculiarity of the Italian context and the time sequence of events creates a favorable condition for carrying out a controlled experiment.
This paper extends to all HSS, with the exception of economics and business, the analysis initiated by Ferrara & Bonaccorsi (2015) on journals in the area of philosophy and history. In the following, we first introduce the database used for the analysis and hence we test for the influence of the journal class on the article score. Some consideration on the results obtained will conclude the paper.
The paper is based on a dataset including data on all the journal articles submitted for evaluation by Italian scholars in the disciplinary areas of architecture, arts and humanities, history and philosophy, law and sociology and political science. Submissions for evaluation took place within the framework of VQR 2004–10, Italy’s national research assessment exercise involving all professors and researchers affiliated to the Italian universities and Public Research Organizations (PROs) as of November 2011. According to adopted rules, research evaluation in HSS was entirely based on peer review; research quality was assessed against the criteria of relevance, meaning contribution to the advancement of the state of art in the field, also in terms of adequacy, efficacy, timeliness and duration of impacts; originality and innovation, meaning contribution to the creation of new knowledge in the field; internationalization, meaning position in the international research landscape. Evaluation has been conducted by five Groups of Evaluation Experts (GEV in the Italian acronym), one for each area in HSS (Architecture; Arts and Humanities; History and Philosophy; Law; Sociology and political sciences); reviewers were instructed by GEV to evaluate articles only on the basis or their merit, regardless of the journal in which they are published in and of the language of publication. Each article had a possible rating of Excellent (A), Good (B), Fair (C) or Limited (D); to each class corresponded a score ranging from 1 (for articles A-rated) to zero (for articles deemed as limited). Negative scores were also assigned in case the article was deemed as non-academic (-1) or for plagiarism or fraud (-2, see Ancaiani et al., 2015 for details). Limited to the human and social sciences, a substantial fraction of articles – namely, 6,701 out of 11,660 (Table 1) – appeared on journals deemed as ‘A-class’ according to the procedure of ASN, intended to select the best researchers for the ranks of associate and full professors. Those journals, according to the relevant Ministerial Decree (No. 76/2012), were those ‘internationally recognized as excellent because of the rigor of their procedures of peer review and because of their diffusion among, esteem by, and impact on, the scholarly community of a field, as indicated also by their presence in the major national and international databases’ (our translation). Most of the remaining articles appeared on journals deemed as ‘academic’ for the purposes of the ASN, while a minority were published in journals that remained ‘uncategorized’. The main feature of the dataset, thus, is that it allows the comparison between the evaluations of journals and individual articles.
A preliminary analysis shows that there is a relationship between the evaluation of individual articles and that of journals where the article is published (Table 2). The non-parametric test for categorical data (Pearson χ2) is statistically significant at 1% (All the statistical analyses have been performed using the software STATA ver. 13 (http://www.stata.com/stata13/)), showing that the two distributions are not independent and hence the two ratings are mutually related. In the following, we will analyze more thoroughly this relationship, also controlling for a number of author-level and article-level variables.
We assume that the probability for an article i, published in the journal j, of receiving a score equal to x ∈ {-2; 1} is influenced by the class assigned to the journal, once controlling for a number of characteristics of the article:
P (Scorei,j = x) = F(Journal classi,j, Paper characteristicsi,j) (1)
Among the controls, we consider the language of publication (Italian or not) and the age (distinguishing among 3 age classes, less than 40 years, between 41 and 55 years and more than 55 years), scientific sector of activity (Scientific Areas 8, 10, 11, 12, 14), academic status (full professor; associate professor; researcher; other) and gender of the researcher. We also add the consideration of two binary variables controlling for the existence of international co-author(s) and for the nationality of the referees (allowing for the possibility of international referees). We finally add a variable taking into account the size of the scientific area of the author. The model is estimated as an ordered probit, an extension of the standard binary probit model, used when the dependent variable takes the form of a ranked and multiple discrete variable, considering alternatively the whole sample or each scientific area; in the first case, we also control for possible area-specific effects. In order to avoid the “dummy trap”, we normalize with respect to articles written in Italian with no international co-author, evaluated by an Italian reviewer, presented by a female researcher in sociology and political science, aged less than 40: i.e. the statistical significance, sign and magnitude of estimated parameters are to be interpreted as differentials with respect to this control group. The total number of available observations amounts to 11,660 varying from a minimum of 918 in architecture to a maximum of 3,838 in law (Table 3).
The main result is that both at the aggregate level and in each scientific area the article score is higher as the journal ranking gets better: in other words, the probability of receiving a high score grows if the article is published in a high-ranking journal according to the evaluation of the ASN’s experts. As for the control variables, we confirm most of the results already emerged in a previous paper on the same data (Cicero et al., 2014), namely, that article scores are higher for papers not written in Italian, with international co-authors, published by an under-40, male full or associate professor. Moreover, we also find that at aggregate level and in most areas an international reviewer and a lower number of professors in the specific scientific sector (SSD) are associated with an higher article score: a possible interpretation of the first result is that the expert groups responsible for the evaluation (GEV) mostly assign to international reviewers more internationalized papers, that are considered to have an higher probability of receiving a high score, given also that the level of internationalization was one of the evaluation criteria according to VQR rules (see again Ancaiani et al., 2015). As for a negative relationship among area size and article score, this result emerged already in Ferrara & Bonaccorsi (2015) for the scientific fields in history and philosophy and is now extended to all HSS: a possible interpretation is that small fields may be favored by a “proximity bias” among authors and reviewers, thus resulting, ceteris paribus, in higher article scores.
As a final check, we concentrate on the probability of receiving an excellent score and relate it to the fact that the article is published in a top, A-Class journal, once controlling for the same variables considered in model 1:
P (Scorei,j = “E”) = F(Journal classi,j = “A”, Paper characteristicsi,j) (2)
In (2), F is the logistic function and the model is estimated as a logit, a class of models allowing to predict the binary response based on the specified predictors. A desirable feature of the logit model is that the regression coefficients may easily be transformed in odds ratio, expressing the change in the odds of the occurrence under scrutiny (in our case, the odds for a paper of receiving an ‘Excellent’ evaluation) due to a small change of a given predictor: in our case, we are particularly interested in the odds associated with the classification of a journal as a top, Class A journal. Estimation results for both the aggregate sample and each scientific area are presented in Table 4.
According to logit estimations, the probability of receiving an excellent evaluation is positively affected by the journal in which the paper is published in: more specifically, publishing in a class A journal almost doubles the probability of receiving an excellent evaluation. Looking at the results in each scientific area, the odds of receiving an excellent evaluation are more than doubled by the publication in a Class A journal in architecture and history and philosophy; the effect is somewhat lower, but still highly significant, in law, and arts and humanities, while disappearing in sociology and political sciences. Logit estimation also broadly confirms the results already emerging from the ordered probit model: the odds of receiving an excellent evaluation are increased by publishing in a foreign language, with an international co-author (albeit only in law and architecture) and when the submitting author is 40 years old or younger, an associate or full professor and a male. Gender effect is in fact significant at the aggregate level and in architecture and humanities, but not in the remaining areas. Also in this case, having an international reviewer and publishing in a SSD characterized by a lower number of full professors helps in obtaining an excellent evaluation.
Using a very large dataset of journal articles published in HSS, the paper proves that independent classifications of journals may be considered as good predictors of the score assigned to individual articles. More specifically, we find that, after controlling for a number of articles’ characteristics, the probability of receiving a better score grows with the quality profile of the journal the article is published in; moreover, the probability of receiving an excellent score almost doubles when the paper is published in a top, A-Class journal. The findings hold both at the aggregate level and for each specific sub-area considered in the analysis. While peer review has to remain the main evaluation methodology, our results indicate that journal classifications may be considered as a useful supporting tool in large evaluation exercise, since it may provide reviewers with valuable information apt to support expert evaluation.
The authors hold the view that it is important to allow the free access to data used in the article in order to enable others to replicate the study. However, information used in the article were gathered by the national agency responsible for evaluation of the University and research system in Italy (ANVUR), in the framework of this VQR exercise. In this context, ANVUR asked Italian professors to provide access to their publications, assuming the commitment not to disclose to the public, unless in an aggregate form, any data concerning the publications submitted for the evaluation and, most importantly, the results of the evaluation itself. This is deemed as necessary in order to guarantee the full anonymity of evaluations performed on each individual publications and on each Italian professor. For this reason, as the public agency in charge of evaluating research of Italian universities, ANVUR does not allow to make information about individual evaluations available to the general public.
The information used to generate data in this article concerning journal classification is available to the public at the following URL: http://www.anvur.it/index.php?option=com_content&view=article&id=254&Itemid=315&lang=it.
The paper is the result of a common effort of the authors. However, Andrea Bonaccorsi can be credited for the “Introduction” and the “Conclusions”, while Antonio Ferrara took care of the “Methods” section and Tindaro Cicero and Marco Malgarini were jointly responsible for the estimates contained in the “The influence of journal classification on the article score” section. All authors have read and agreed to the final content of the manuscript.
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
Competing Interests: No competing interests were disclosed.
Competing Interests: No competing interests were disclosed.
Competing Interests: No competing interests were disclosed.
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | |||
---|---|---|---|
1 | 2 | 3 | |
Version 1 07 Jul 15 |
read | read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
“On the one hand, a panel of experts classified […] academic journals in A-rated and non A-rated. The rating exercise was done on the basis of the reputation, esteem, diffusion and impact of journals, that is, on a qualitative, expert-based, reputational basis. On the other hand, we also have the rating of individual articles published in those journals, which have been done by a large number of individual referees (not panels) and summarized with a consensus agreement approach by expert panels, who however acted independently from the other panels, and without exchange of information. This peculiarity of the Italian context and the time sequence of events creates a favorable condition for carrying out a controlled experiment.”
Bonaccorsi et al. build a dataset by matching two different administrative databases produced, by the Italian National Agency for the Evaluation of Universities and Research Institutes (ANVUR) in their institutional activity.[2] In this comment I will argue that according to ANVUR official records,[3] the “two evaluations” cannot be considered as being “independent”, because they were reached through two intertwined procedures. As a consequence, the positive relationship among peer review evaluations and journal rating found by Bonaccorsi et al. cannot be considered as a sound evidence that “journal ratings are good predictors of article quality”.
1. Consider first the evaluations of individual articles. It was developed in the context of the national research assessment exercise covering the years 2004-2010 (hereinafter indicated by its Italian acronym, VQR). The peer review process started in September 2012 and concluded in February-March 2013. It was “conducted by five Groups of Evaluation Experts (GEV in the Italian acronym), one for each [research] area”. The members of the five GEVs were nominated by the board of ANVUR,[4] an issue that will become relevant later on.
GEV members were in charge of choosing reviewers and to define evaluation criteria by considering the general ones stated by the ANVUR board (more detailed descriptions of the procedures are available in Baccini and De Nicolao 2016, Ancaiani et al. 2015). However, the first accomplished duty of all the five GEVs was the development of “rankings of journals”, by classifying journals in two or three classes. These rankings[5] were published by March 2012, before the starting of the peer review process.[6]
The peer review process was described in official documents as an “informed peer review” process,[7] because each article was evaluated by two reviewers knowing the complete metadata of the article, in particular the journal in which it was published. They knew also the rank of that journal according to the ranking developed by the GEV that called them as reviewers.[8] Bonaccorsi et al. wrote:
“reviewers were instructed by GEV to evaluate articles only on the basis of their merit regardless of the journal in which they are published in and of the language of publication”.
This does not correspond to the official records of the procedure. Indeed, GEV8 for Architecture “produced and used in the informed peer review the journal ranking”.[9] The GEV11 for History, philosophy, pedagogy and psychology asked reviewers to “consider also … the quality of publication venue”. Reviewers were furnished with the “journal rankings, which, however, they were not obliged to comply”. Moreover, for articles published in journal classified as A, GEV11 might have decided also to ask only one reviewer report.[10] The GEV12 for law, explicitly stated that its journal ranking had to be considered as one source of information among the others available to reviewers for their evaluation.[11] And analogously the GEV14 for Political and social sciences considered the ranking as a way to furnish to the reviewers, “especially [to] foreign reviewers” “better information (according to the logic of informed peer review)”.[12] Only the GEV10 for antiquities, philology, literary studies, art history, established that reviewers, even though they knew the published journal ranking, had to perform a “peer review” where “the venue of publishing, the type and the language in which the research has been expressed are not factors conditioning the evaluation”.[13]
Last but not least: if two reviewers disagreed about the evaluation of an article, a consensus group decided the final evaluation. Consensus groups were generally composed by two members of the GEV, who ranked also journals. Data about the number of journal articles evaluated by consensus groups were not available (on this point see Baccini and De Nicolao 2016).
In short: we can affirm that journal rankings developed by GEVs were used by reviewers as information for evaluating articles, according to “the logic of informed peer review”. Readers have to wait a bit for understanding the relevance of this statement.. Consider now the journal ratings used by Bonaccorsi et al.. It was developed in the context of the National Scientific Habilitation (hereinafter ASN, according to Italian acronym). According to the Ministerial decree[14] ruling the ASN, ANVUR was charged of producing a list of scientific journals and a list of class A journals.[15]
2. For developing the journal lists, “the working group on books and scientific journals”, coordinated by Anvur president, was constituted. It was composed by the members of the ANVUR board, and by other scholars nominated by the ANVUR board and organized in five different Panels,[16] defined in reference to the research areas already described for the GEVs . Each panel was composed by 4 members, with the only exception of Panel10 composed by 8 members.[17] Recall that ANVUR board had already nominated the GEVs. Works and criteria adopted by the Panels for the ratings were only summarily described in short final reports, with the exception of Panel10 that never published a report for its work.[18]
The ministerial decree stated that for developing the journal lists, ANVUR “avails upon the GEVs for the VQR”.[19] According to this provision, ANVUR stated that it “retains to get opinions from GEV”[20] and that it will deliver the final lists of journals after having considered “the observations and proposals of the GEVs”.[21]
5th October 2012, some weeks after the definitive publication of journal lists, and for replying to a growing mass of criticisms (Mazzotti 2012, Baccini 2016), ANVUR published a document where it was clearly stated that Class A lists developed by the Panels, “as suggested by the Ministerial Decree 76/2012, were sent to the GEVs by asking their opinions”. The only exception was for GEV10. The opinions expressed by the GEVs were positive, with “small modifications” required by GEV11 and by GEV14.[22]
According to available official records, GEVs not only gave their final opinion, but interacted in many ways with panels.[23] These are the interactions documented in the publicly available records:
3. Consider now the “time sequence of events”. The journal rankings for the VQR developed by the GEVs were ready by March 2012; the journal ratings developed by Panels for the ASN were ready by July-August 2012; the informed peer review began in late September 2012. Reviewers called to evaluate papers for the VQR, at least Italian reviewers, knew that journal ratings were developed for the ASN, and that they were only partially different from the ones developed by the GEVs for the VQR. Reviewers, “in the logic of informed peer review” might have used the ratings for the ASN as information for their evaluations of articles. This short circuit was testified by GEV10 by explicitly writing in its final report that the publication of the class A list was an “element of disruption” for the VQR. [30]
This is another reasons for which the two evaluations used by the Bonaccorsi et al. cannot be considered as “independent”.
4. This comment documents that according to available official records, data used by Bonaccorsi et al. about journal ratings and scores of individual articles cannot be considered as generated by “two independent evaluations”, whatever the meaning of the adjective “independent”. Indeed the two evaluations were intertwined in various ways. They were organized by a same group of scholars: the ANVUR board that was in charge also of choosing the members of both the GEVs and the Panels. ANVUR board, GEVs and Panels interacted during the development of ASN journal ratings; these ratings were developed starting by the VQR journal rankings developed by GEVs. Finally reviewers, in charge of evaluating articles for the VQR, were asked to consider the venues of publication and in particular the journal rankings. It is therefore hardly surprising that Bonaccorsi et al. found that “the probability [of articles] of receiving a better scores grows with the quality profile of the journal”. Bonaccorsi et al.’s paper did not represent a sound addition to the body of literature on journal classification.
References
Ancaiani, Alessio, Alberto F. Anfossi, Anna Barbara, Sergio Benedetto, Brigida Blasi, Valentina Carletti, Tindaro Cicero, Alberto Ciolfi, Filippo Costa, Giovanna Colizza, Marco Costantini, Fabio di Cristina, Antonio Ferrara, Rosa M. Lacatena, Marco Malgarini, Irene Mazzotta, Carmela A. Nappi, Sandra Romagnosi, and Serena Sileoni. 2015. "Evaluating scientific research in Italy: The 2004–10 research evaluation exercise." Research Evaluation no. 24 (3):242-255. doi: 10.1093/reseval/rvv008.
Baccini, A. 2016. "Napoléon et l’évaluation bibliométrique de la recherche. Considérations sur la réforme de l’université et sur l’action de l’agence national d’évaluation en Italie." Canadian Journal of Information and Library Science-Revue Canadienne des Sciences de l'Information et de Bibliotheconomie no. 40 (1):37-57. doi: 10.1353/ils.2016.0003.
Baccini, Alberto, and Giuseppe De Nicolao. 2016. "Do they agree? Bibliometric evaluation versus informed peer review in the Italian research assessment exercise." Scientometrics:1-21. doi: 10.1007/s11192-016-1929-y.
Bonaccorsi, A, T Cicero, A Ferrara, and M Malgarini. 2015. Journal ratings as predictors of articles quality in Arts, Humanities and Social Sciences: an analysis based on the Italian Research Evaluation Exercise [version 1; referees: 3 approved]. Vol. 4.
Mazzotti, Massimo. 2012. Listing wildly. Times Higher Education, 08/11.
NOTES
[1] Raw data were not disclosed to other scholars, not even in a feasible anonymized form. The link provided in the section “Data availability” of the paper does not contain any more data concerning the article. I will provide below the direct links to the relevant documents contained in the ANVUR website. In many cases I found the links by using Wayback machine https://archive.org/web/.
[2] Bonaccorsi et al. inappropriately described that research activity as a “controlled experiment”.
[3] In this comment I will provide my own English translations for quotations.
[4] Andrea Bonaccorsi was one of the seven members of the board of ANVUR.
[5] Hereinafter the words “ranking/rankings” refers only to the journal ranking developed by GEVs for VQR, and the words “rating/ratings” refers only to the journal classification developed by Panels for National Scientific Habilitation and used by Bonaccorsi et al..
[6] http://www.anvur.org/rapporto/main.php?paragraph=3.1.1&cap=My4xLjEuIFtlbV1MYSBjbGFzc2lmaWNhemlvbmUgZGVsbGUgcml2aXN0ZVsvZW1d
[7] http://www.anvur.org/rapporto/files/VQR2004-2010_RapportoFinale_parteprima.pdf, p. 5 and passim.
[8] http://www.anvur.org/rapporto/files/VQR2004-2010_RapportoFinale_parteprima.pdf, p. 21.
[9] ibidem
[10] http://www.anvur.org/rapporto/files/Area11/VQR2004-2010_Area11_RapportoFinale.pdf, pp. 33-34.
[11] http://www.anvur.org/rapporto/files/Area12/VQR2004-2010_Area12_RapportoFinale.pdf, pp. 32-34. And also http://www.anvur.org/rapporto/files/Area12/VQR2004-2010_Area12_Appendici.pdf
[12] http://www.anvur.org/rapporto/files/Area14/VQR2004-2010_Area14_RapportoFinale.pdf, p. 17.
[13] http://www.anvur.org/rapporto/files/Area10/VQR2004-2010_Area10_Appendici.pdf. Appendice B, p. 7. The effectiveness of this recommendation cannot be verified.
[14] Ministerial decree n. 76/2012, art. 6.6. http://attiministeriali.miur.it/anno-2012/giugno/dm-07062012.aspx
[15] Annex b of the Ministerial Decree, http://attiministeriali.miur.it/media/192907/dm_07_06_2012_allegatob.pdf. ANVUR stated its duties in the deliberation n. 50/2012; http://www.anvur.org/attachments/article/252/Delibera50_12.pdf
[16] This provision permitted to the members of the ANVUR board to participate to the meetings of the panels. Since the minutes of the meetings were not published, it is impossible to know if ANVUR board members really participated to the meetings and how they interacted with the other panelists in the final decisions.
[17] The composition of the panels was defined in these ANVUR deliberations: http://www.anvur.org/attachments/article/422/delibera55_12.pdf; http://www.anvur.org/attachments/article/422/delibera58_12.pdf; http://www.anvur.org/attachments/article/422/delibera63_12.pdf.
[18] http://www.anvur.org/attachments/article/254/Relazionefinale_GdLArea08.pdf ; http://www.anvur.org/attachments/article/254/Relazionefinale_GdLArea11.pdf; http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea12.pdf; http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea14.pdf.
[19] http://attiministeriali.miur.it/anno-2012/giugno/dm-07062012.aspx, Annex B, section 2.
[20] http://www.anvur.org/attachments/article/252/Delibera50_12.pdf , Art. 11.6.
[21] Ibidem, art 12.3.
[22] GEV14’s request of modification was limited to the sub-field of political sciences http://www.roars.it/online/wp-content/uploads/2012/10/chiarimenti_riviste_scientifiche.pdf, pp. 1-2. On 24th September 2012, ANVUR published a document, where it was clearly stated that “the lists were previously submitted to the opinion of the GEV”. After 4 days the document was replaced by another one. The only difference between the two documents was that the sentence quoted above was dropped in the second document. Copy of the documents and a discussion is available here: http://www.roars.it/online/lenigmistica-di-anvur-trovate-le-differenze/
[23] In an ANVUR document of January 2013, it was clearly stated that ANVUR developed the journal lists “by employing … the GEVs and the working group [on books and scientific journals]. The document was an undated pdf. The properties of the document registered it as last modified on 13th January 2013, and as authored by “Bonaccorsi” http://www.anvur.org/attachments/article/252/riviste.pdf, p. 2.
[24] http://www.roars.it/online/wp-content/uploads/2012/09/chiarimenti_riviste_classea_0.pdf
[25] Art. 12.4 of http://www.anvur.org/attachments/article/252/Delibera50_12.pdf.
[26] For an example of such a letter: http://www.glottologia.org/wp-content/uploads/2012/07/Documento-10-1-ANVUR.pdf.
[27] The complete description is available here: http://www.anvur.org/attachments/article/252/riviste.pdf
[28] http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea14.pdf, p. 1
[29] The complete description is available here: http://www.anvur.org/attachments/article/252/riviste.pdf.
[30] http://www.anvur.org/rapporto/files/Area10/VQR2004-2010_Area10_RapportoFinale.pdf, p. 26.
“On the one hand, a panel of experts classified […] academic journals in A-rated and non A-rated. The rating exercise was done on the basis of the reputation, esteem, diffusion and impact of journals, that is, on a qualitative, expert-based, reputational basis. On the other hand, we also have the rating of individual articles published in those journals, which have been done by a large number of individual referees (not panels) and summarized with a consensus agreement approach by expert panels, who however acted independently from the other panels, and without exchange of information. This peculiarity of the Italian context and the time sequence of events creates a favorable condition for carrying out a controlled experiment.”
Bonaccorsi et al. build a dataset by matching two different administrative databases produced, by the Italian National Agency for the Evaluation of Universities and Research Institutes (ANVUR) in their institutional activity.[2] In this comment I will argue that according to ANVUR official records,[3] the “two evaluations” cannot be considered as being “independent”, because they were reached through two intertwined procedures. As a consequence, the positive relationship among peer review evaluations and journal rating found by Bonaccorsi et al. cannot be considered as a sound evidence that “journal ratings are good predictors of article quality”.
1. Consider first the evaluations of individual articles. It was developed in the context of the national research assessment exercise covering the years 2004-2010 (hereinafter indicated by its Italian acronym, VQR). The peer review process started in September 2012 and concluded in February-March 2013. It was “conducted by five Groups of Evaluation Experts (GEV in the Italian acronym), one for each [research] area”. The members of the five GEVs were nominated by the board of ANVUR,[4] an issue that will become relevant later on.
GEV members were in charge of choosing reviewers and to define evaluation criteria by considering the general ones stated by the ANVUR board (more detailed descriptions of the procedures are available in Baccini and De Nicolao 2016, Ancaiani et al. 2015). However, the first accomplished duty of all the five GEVs was the development of “rankings of journals”, by classifying journals in two or three classes. These rankings[5] were published by March 2012, before the starting of the peer review process.[6]
The peer review process was described in official documents as an “informed peer review” process,[7] because each article was evaluated by two reviewers knowing the complete metadata of the article, in particular the journal in which it was published. They knew also the rank of that journal according to the ranking developed by the GEV that called them as reviewers.[8] Bonaccorsi et al. wrote:
“reviewers were instructed by GEV to evaluate articles only on the basis of their merit regardless of the journal in which they are published in and of the language of publication”.
This does not correspond to the official records of the procedure. Indeed, GEV8 for Architecture “produced and used in the informed peer review the journal ranking”.[9] The GEV11 for History, philosophy, pedagogy and psychology asked reviewers to “consider also … the quality of publication venue”. Reviewers were furnished with the “journal rankings, which, however, they were not obliged to comply”. Moreover, for articles published in journal classified as A, GEV11 might have decided also to ask only one reviewer report.[10] The GEV12 for law, explicitly stated that its journal ranking had to be considered as one source of information among the others available to reviewers for their evaluation.[11] And analogously the GEV14 for Political and social sciences considered the ranking as a way to furnish to the reviewers, “especially [to] foreign reviewers” “better information (according to the logic of informed peer review)”.[12] Only the GEV10 for antiquities, philology, literary studies, art history, established that reviewers, even though they knew the published journal ranking, had to perform a “peer review” where “the venue of publishing, the type and the language in which the research has been expressed are not factors conditioning the evaluation”.[13]
Last but not least: if two reviewers disagreed about the evaluation of an article, a consensus group decided the final evaluation. Consensus groups were generally composed by two members of the GEV, who ranked also journals. Data about the number of journal articles evaluated by consensus groups were not available (on this point see Baccini and De Nicolao 2016).
In short: we can affirm that journal rankings developed by GEVs were used by reviewers as information for evaluating articles, according to “the logic of informed peer review”. Readers have to wait a bit for understanding the relevance of this statement.. Consider now the journal ratings used by Bonaccorsi et al.. It was developed in the context of the National Scientific Habilitation (hereinafter ASN, according to Italian acronym). According to the Ministerial decree[14] ruling the ASN, ANVUR was charged of producing a list of scientific journals and a list of class A journals.[15]
2. For developing the journal lists, “the working group on books and scientific journals”, coordinated by Anvur president, was constituted. It was composed by the members of the ANVUR board, and by other scholars nominated by the ANVUR board and organized in five different Panels,[16] defined in reference to the research areas already described for the GEVs . Each panel was composed by 4 members, with the only exception of Panel10 composed by 8 members.[17] Recall that ANVUR board had already nominated the GEVs. Works and criteria adopted by the Panels for the ratings were only summarily described in short final reports, with the exception of Panel10 that never published a report for its work.[18]
The ministerial decree stated that for developing the journal lists, ANVUR “avails upon the GEVs for the VQR”.[19] According to this provision, ANVUR stated that it “retains to get opinions from GEV”[20] and that it will deliver the final lists of journals after having considered “the observations and proposals of the GEVs”.[21]
5th October 2012, some weeks after the definitive publication of journal lists, and for replying to a growing mass of criticisms (Mazzotti 2012, Baccini 2016), ANVUR published a document where it was clearly stated that Class A lists developed by the Panels, “as suggested by the Ministerial Decree 76/2012, were sent to the GEVs by asking their opinions”. The only exception was for GEV10. The opinions expressed by the GEVs were positive, with “small modifications” required by GEV11 and by GEV14.[22]
According to available official records, GEVs not only gave their final opinion, but interacted in many ways with panels.[23] These are the interactions documented in the publicly available records:
3. Consider now the “time sequence of events”. The journal rankings for the VQR developed by the GEVs were ready by March 2012; the journal ratings developed by Panels for the ASN were ready by July-August 2012; the informed peer review began in late September 2012. Reviewers called to evaluate papers for the VQR, at least Italian reviewers, knew that journal ratings were developed for the ASN, and that they were only partially different from the ones developed by the GEVs for the VQR. Reviewers, “in the logic of informed peer review” might have used the ratings for the ASN as information for their evaluations of articles. This short circuit was testified by GEV10 by explicitly writing in its final report that the publication of the class A list was an “element of disruption” for the VQR. [30]
This is another reasons for which the two evaluations used by the Bonaccorsi et al. cannot be considered as “independent”.
4. This comment documents that according to available official records, data used by Bonaccorsi et al. about journal ratings and scores of individual articles cannot be considered as generated by “two independent evaluations”, whatever the meaning of the adjective “independent”. Indeed the two evaluations were intertwined in various ways. They were organized by a same group of scholars: the ANVUR board that was in charge also of choosing the members of both the GEVs and the Panels. ANVUR board, GEVs and Panels interacted during the development of ASN journal ratings; these ratings were developed starting by the VQR journal rankings developed by GEVs. Finally reviewers, in charge of evaluating articles for the VQR, were asked to consider the venues of publication and in particular the journal rankings. It is therefore hardly surprising that Bonaccorsi et al. found that “the probability [of articles] of receiving a better scores grows with the quality profile of the journal”. Bonaccorsi et al.’s paper did not represent a sound addition to the body of literature on journal classification.
References
Ancaiani, Alessio, Alberto F. Anfossi, Anna Barbara, Sergio Benedetto, Brigida Blasi, Valentina Carletti, Tindaro Cicero, Alberto Ciolfi, Filippo Costa, Giovanna Colizza, Marco Costantini, Fabio di Cristina, Antonio Ferrara, Rosa M. Lacatena, Marco Malgarini, Irene Mazzotta, Carmela A. Nappi, Sandra Romagnosi, and Serena Sileoni. 2015. "Evaluating scientific research in Italy: The 2004–10 research evaluation exercise." Research Evaluation no. 24 (3):242-255. doi: 10.1093/reseval/rvv008.
Baccini, A. 2016. "Napoléon et l’évaluation bibliométrique de la recherche. Considérations sur la réforme de l’université et sur l’action de l’agence national d’évaluation en Italie." Canadian Journal of Information and Library Science-Revue Canadienne des Sciences de l'Information et de Bibliotheconomie no. 40 (1):37-57. doi: 10.1353/ils.2016.0003.
Baccini, Alberto, and Giuseppe De Nicolao. 2016. "Do they agree? Bibliometric evaluation versus informed peer review in the Italian research assessment exercise." Scientometrics:1-21. doi: 10.1007/s11192-016-1929-y.
Bonaccorsi, A, T Cicero, A Ferrara, and M Malgarini. 2015. Journal ratings as predictors of articles quality in Arts, Humanities and Social Sciences: an analysis based on the Italian Research Evaluation Exercise [version 1; referees: 3 approved]. Vol. 4.
Mazzotti, Massimo. 2012. Listing wildly. Times Higher Education, 08/11.
NOTES
[1] Raw data were not disclosed to other scholars, not even in a feasible anonymized form. The link provided in the section “Data availability” of the paper does not contain any more data concerning the article. I will provide below the direct links to the relevant documents contained in the ANVUR website. In many cases I found the links by using Wayback machine https://archive.org/web/.
[2] Bonaccorsi et al. inappropriately described that research activity as a “controlled experiment”.
[3] In this comment I will provide my own English translations for quotations.
[4] Andrea Bonaccorsi was one of the seven members of the board of ANVUR.
[5] Hereinafter the words “ranking/rankings” refers only to the journal ranking developed by GEVs for VQR, and the words “rating/ratings” refers only to the journal classification developed by Panels for National Scientific Habilitation and used by Bonaccorsi et al..
[6] http://www.anvur.org/rapporto/main.php?paragraph=3.1.1&cap=My4xLjEuIFtlbV1MYSBjbGFzc2lmaWNhemlvbmUgZGVsbGUgcml2aXN0ZVsvZW1d
[7] http://www.anvur.org/rapporto/files/VQR2004-2010_RapportoFinale_parteprima.pdf, p. 5 and passim.
[8] http://www.anvur.org/rapporto/files/VQR2004-2010_RapportoFinale_parteprima.pdf, p. 21.
[9] ibidem
[10] http://www.anvur.org/rapporto/files/Area11/VQR2004-2010_Area11_RapportoFinale.pdf, pp. 33-34.
[11] http://www.anvur.org/rapporto/files/Area12/VQR2004-2010_Area12_RapportoFinale.pdf, pp. 32-34. And also http://www.anvur.org/rapporto/files/Area12/VQR2004-2010_Area12_Appendici.pdf
[12] http://www.anvur.org/rapporto/files/Area14/VQR2004-2010_Area14_RapportoFinale.pdf, p. 17.
[13] http://www.anvur.org/rapporto/files/Area10/VQR2004-2010_Area10_Appendici.pdf. Appendice B, p. 7. The effectiveness of this recommendation cannot be verified.
[14] Ministerial decree n. 76/2012, art. 6.6. http://attiministeriali.miur.it/anno-2012/giugno/dm-07062012.aspx
[15] Annex b of the Ministerial Decree, http://attiministeriali.miur.it/media/192907/dm_07_06_2012_allegatob.pdf. ANVUR stated its duties in the deliberation n. 50/2012; http://www.anvur.org/attachments/article/252/Delibera50_12.pdf
[16] This provision permitted to the members of the ANVUR board to participate to the meetings of the panels. Since the minutes of the meetings were not published, it is impossible to know if ANVUR board members really participated to the meetings and how they interacted with the other panelists in the final decisions.
[17] The composition of the panels was defined in these ANVUR deliberations: http://www.anvur.org/attachments/article/422/delibera55_12.pdf; http://www.anvur.org/attachments/article/422/delibera58_12.pdf; http://www.anvur.org/attachments/article/422/delibera63_12.pdf.
[18] http://www.anvur.org/attachments/article/254/Relazionefinale_GdLArea08.pdf ; http://www.anvur.org/attachments/article/254/Relazionefinale_GdLArea11.pdf; http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea12.pdf; http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea14.pdf.
[19] http://attiministeriali.miur.it/anno-2012/giugno/dm-07062012.aspx, Annex B, section 2.
[20] http://www.anvur.org/attachments/article/252/Delibera50_12.pdf , Art. 11.6.
[21] Ibidem, art 12.3.
[22] GEV14’s request of modification was limited to the sub-field of political sciences http://www.roars.it/online/wp-content/uploads/2012/10/chiarimenti_riviste_scientifiche.pdf, pp. 1-2. On 24th September 2012, ANVUR published a document, where it was clearly stated that “the lists were previously submitted to the opinion of the GEV”. After 4 days the document was replaced by another one. The only difference between the two documents was that the sentence quoted above was dropped in the second document. Copy of the documents and a discussion is available here: http://www.roars.it/online/lenigmistica-di-anvur-trovate-le-differenze/
[23] In an ANVUR document of January 2013, it was clearly stated that ANVUR developed the journal lists “by employing … the GEVs and the working group [on books and scientific journals]. The document was an undated pdf. The properties of the document registered it as last modified on 13th January 2013, and as authored by “Bonaccorsi” http://www.anvur.org/attachments/article/252/riviste.pdf, p. 2.
[24] http://www.roars.it/online/wp-content/uploads/2012/09/chiarimenti_riviste_classea_0.pdf
[25] Art. 12.4 of http://www.anvur.org/attachments/article/252/Delibera50_12.pdf.
[26] For an example of such a letter: http://www.glottologia.org/wp-content/uploads/2012/07/Documento-10-1-ANVUR.pdf.
[27] The complete description is available here: http://www.anvur.org/attachments/article/252/riviste.pdf
[28] http://www.anvur.org/attachments/article/254/relazionefinale_gdlarea14.pdf, p. 1
[29] The complete description is available here: http://www.anvur.org/attachments/article/252/riviste.pdf.
[30] http://www.anvur.org/rapporto/files/Area10/VQR2004-2010_Area10_RapportoFinale.pdf, p. 26.