Do we measure novelty when we analyze unusual combinations of cited references? A validation study of bibliometric novelty indicators based on F1000Prime data

被引:53
作者
Bornmann, Lutz [1 ]
Tekles, Alexander [1 ,2 ]
Zhang, Helena H. [3 ,4 ,5 ]
Ye, Fred Y. [3 ,4 ,5 ]
机构
[1] Max Planck Gesell, Div Sci & Innovat Studies, Adm Headquarters, Hofgartenstr 8, D-80539 Munich, Germany
[2] Ludwig Maximilians Univ Munchen, Dept Sociol, Konradstr 6, D-80801 Munich, Germany
[3] Nanjing Univ, Sch Informat Management, Jiangsu Key Lab Data Engn & Knowledge Serv, Nanjing 210023, Jiangsu, Peoples R China
[4] Nanjing Univ, IJIL, Nanjing, Jiangsu, Peoples R China
[5] Univ Illinois, IJIL, Champaign, IL 61820 USA
基金
中国国家自然科学基金;
关键词
Bibliometrics; Novelty; Creativity; Cited references; F1000Prime; CONVERGENT VALIDITY; CREATIVITY; IMPACT; PUBLICATIONS; SCIENCE;
D O I
10.1016/j.joi.2019.100979
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Lee et al. (2015) - based on Uzzi et al. (2013) - and Wang et al. (2017) proposed scores based on cited references (cited journals) data which can be used to measure the novelty of papers (named as novelty scores U and Win this study). Although previous research has used novelty scores in various empirical analyses, no study has been published up to now to the best of our knowledge - which quantitatively tested the convergent validity of novelty scores: do these scores measure what they propose to measure? Using novelty assessments by faculty members (FMs) at F1000Prime for comparison, we tested the convergent validity of the two novelty scores (U and W). FMs' assessments do not only refer to the quality of biomedical papers, but also to their characteristics (by assigning certain tags to the papers): for example, are the presented findings or formulated hypotheses novel (tags "new findings" and "hypothesis")? We used these and other tags to investigate the convergent validity of both novelty scores. Our study reveals different results for the novelty scores: the results for novelty score U are mostly in agreement with previously formulated expectations. We found, for instance, that for a standard deviation (one unit) increase in novelty score U, the expected number of assignments of the "new finding" tag increase by 7.47%. The results for novelty score W, however, do not reflect convergent validity with the FMs' assessments: only the results for some tags are in agreement with the expectations. Thus, we propose based on our results - the use of novelty score U for measuring novelty quantitatively, but question the use of novelty score W. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页数:15
相关论文
共 59 条
[1]  
Acock, 2018, GENTLE INTRO STATA
[2]  
Adams J., 2016, Interdisciplinary research: Methodologies for identification and assessment
[3]  
Angeles G., 2014, A guide to longitudinal program impact evaluation
[4]   Revolutionizing peer review? [J].
不详 .
NATURE NEUROSCIENCE, 2005, 8 (04) :397-397
[5]  
[Anonymous], 2017, Web indicators for research evaluation: A practical guide
[6]  
Belter CW, 2018, CHANDOS INF PROF SER, P33, DOI 10.1016/B978-0-08-102017-3.00004-8
[7]  
Berger JM., 2014, Rajiv Gandhi Univ Health Sci J Pharmaceut Sci, V4, P81
[8]   Heuristics as conceptual lens for understanding and studying the usage of bibliometrics in research evaluation [J].
Bornmann, Lutz ;
Marewski, Julian N. .
SCIENTOMETRICS, 2019, 120 (02) :419-459
[9]   How well does I3 perform for impact measurement compared to other bibliometric indicators? The convergent validity of several (field-normalized) indicators [J].
Bornmann, Lutz ;
Tekles, Alexander ;
Leydesdorff, Loet .
SCIENTOMETRICS, 2019, 119 (02) :1187-1205
[10]   Interrater reliability and convergent validity of F1000Prime peer review [J].
Bornmann, Lutz .
JOURNAL OF THE ASSOCIATION FOR INFORMATION SCIENCE AND TECHNOLOGY, 2015, 66 (12) :2415-2426