Hostname: page-component-586b7cd67f-gb8f7 Total loading time: 0 Render date: 2024-11-22T17:31:01.032Z Has data issue: false hasContentIssue false

Using bibliometrics to evaluate translational science training: evidence for early career success of KL2 scholars

Published online by Cambridge University Press:  24 July 2020

Kelli Qua*
Affiliation:
School of Medicine, Clinical and Translational Science Collaborative, Case Western Reserve University, Cleveland, OH, USA
Clara M. Pelfrey
Affiliation:
School of Medicine, Clinical and Translational Science Collaborative, Case Western Reserve University, Cleveland, OH, USA
*
Address for correspondence: K. Qua, PhD, School of Medicine, Clinical and Translational Science Collaborative, Case Western Reserve University, BRB 109, 10900 Euclid Ave., Cleveland, OH44106-4961, USA. Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Introduction:

Evaluating clinical and translational research (CTR) mentored training programs is challenging because no two programs are alike. Careful selection of appropriate metrics is required to make valid comparisons between individuals and between programs. The KL2 program provides mentored-training for early-stage CTR investigators. Clinical and Translational Awards across the country have unique KL2 programs. The evaluation of KL2 programs has begun to incorporate bibliometrics to measure KL2 scholar and program impact.

Methods:

This study investigated demographic differences in bibliometric performance and post-K award funding of KL2 scholars and compared the bibliometric performance and post-K award federal funding of KL2 scholars and other mentored-K awardees at the same institution. Data for this study included SciVal and iCite bibliometrics and National Institutions of Health RePORTER grant information for mentored-K awardees (K08, K23, and KL2) at Case Western Reserve University between 2005 and 2013.

Results:

Results showed no demographics differences within the KL2 program scholars. Bibliometric differences between KL2 and other mentored-K awardee indicated an initial KL2 advantage for the number of publications at 5 years’ post-matriculation (i.e., the start of the K award). Regression analyses indicated the number of initial publications was a significant predictor of federal grant funding at the same time point. Analysis beyond the 5-year post-matriculation point did not result in a sustained, significant KL2 advantage.

Conclusions:

Factors that contributed to the grant funding advantage need to be determined. Additionally, differences between translational and clinical bibliometrics must be interpreted with caution, and appropriate metrics for translational science must be established.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Association for Clinical and Translational Science 2020

Introduction

Rigorous evaluation of a clinical and translational research (CTR) training program requires careful selection of reliable quantitative metrics and matching to a reasonable comparison group. This study expands evaluation methods to assess the KL2 program. The KL2 program is a multiyear mentored training award focused on career development for early-stage CTR investigators. The KL2 program is a workforce development component of the larger Clinical and Translational Award (CSTA) granted by the National Center for Advancing Translational Sciences (NCATS) at the National Institutions of Health (NIH). Over 60 funded CSTAs across the country have KL2 programs and no two are alike. While the programs vary greatly, they do include several overlapping characteristics such as interdisciplinary mentorship, training opportunities aligned to CTR competencies [Reference Calvin-Naylor1] such as team science, scientific and grant writing, career development planning, and protected research time. Outside of these programmatic requirements that are standard across Clinical and Translational Science Awards (CTSAs), one recent study investigating KL2 program differences surveyed 55 CSTA hubs (90%) and reported that 87% of KL2 scholars had 2–3 years of supported (i.e., funded) KL2 training [Reference Sorkness2]; at least one CTSA provides 4 years of funding. The variability in length of funding is just one example of the various KL2 program differences, which make a standardized evaluation difficult.

The extent of program-level differences is also present when considering the outcome measures of the KL2 program across the CSTAs. While the primary outcome across KL2 programs is to establish independent CTR investigators, institutions operationalize “independent funding” differently. Some examples of the variability in the definition of achieving independent funding include: principal investigator (PI) roles on any grant, PI on an NIH R01 grant, other key roles such as Co-PI or Co-I on an NIH R01 grant, R-level funding (either NIH or Non-NIH), and time to R01 [Reference Sweeney3Reference Narahari5]. Independent funding, in any of these forms, may be the primary outcome for KL2 programs, but it is not the singular outcome. When KL2 programs only use independent grant funding as a measure of KL2 program success, they fail to account for the more intermediate outcomes that impact funding such a publications, collaborations, and smaller foundation and/or pilot grant awards, for example.

Recently, the evaluation of KL2 programs has looked to bibliometrics (i.e., the science and quantification of publication data) to provide a broader picture of KL2 scholar and program impact. Several methods for using bibliometrics in this capacity have been tested for their efficacy, such as comparisons of paid services (e.g., Elsevier products) and the use of open-source tools (e.g., iCite) [Reference Sayavedra, Hogle and Moberg6]. The strengths of these bibliometric methods are that they provide insight into a more diverse set of research outcomes, including collaboration, dissemination patterns, citation and publication networks, and publication impact [Reference Llewellyn7]. Advanced bibliometric methods can facilitate the evaluation of unique KL2 programs by expanding the definition of research productivity [Reference Yu8].

In an effort to expand the existing evaluation efforts conducted by other CTSAs, this study aimed to explore the bibliometric outcomes and federal follow-on funding of KL2 scholars in comparison to other K scholars at the same institution in order to assess the impact of the KL2 program on research outcomes. The present research had three objectives: (1) to investigate whether there were demographic differences (e.g., gender, race, and degree) in bibliometric performance or post-K award federal grant funding of KL2 scholars, (2) to compare the bibliometric performance between KL2 scholars and other to other mentored-K awardees (K08 and K23), and (3) to compare post-K award federal grant funding of KL2 scholars compared to other mentored-K awardees (K08 and K23).

Methods

Participants

Early-stage investigators who were awarded NIH clinical, translational, or behavioral mentored-K awards (K08, K23, KL2) at Case Western Reserve University (CWRU) between 2005 and 2013 were included in this analysis. K08 and K23 mechanisms were selected as a comparison group because they are similar to the KL2 mechanism in that (1) these mechanisms provide protected time for early-stage investigators and (2) they focus on scholars in “biomedical and behavioral research, including translational research” and “patient-oriented research” [9]. At CWRU, the K08 and K23 awards differ from the KL2 in that K08 and K23 scholars are not required to take coursework, have different funding lengths (i.e., the KL2 program studied here provides 4 years of funding; the K08/K23 scholars reported between three to 5 years of funding, with an average funding length of 4.5 years), and typically focus on clinical research. Other NIH career development K awards exist (e.g., K01, K25, K99), but these K awards do not specifically target early-stage CTR and/or behavioral investigators.

Participants were selected using the date range of 2005–2013 to ensure all scholars had a minimum of 5 years of data from the start of the K award (i.e., matriculation). Thirteen K12 scholars (NIH institutional career development awards) from 2005 to 2006 were transitioned to KL2 scholars when the institution’s CTSA was initially awarded in 2007. Demographics of the K12s from 2005 to 2006 and KL2 scholars are comparable.

Measures

A series of bibliometrics were investigated in order to produce a valid and reliable data set. One set of metrics was exported from the Elsevier’s SciVal [10] database, and a second set of metrics was exported from the NIH’s iCite [11] database. Extracting information from both of these databases reduces possible bias and provides a more reliable estimate of the quantity and impact of scholarly work. iCite is limited to analyzing only articles that appear in PubMed and is a public and government-verified source, whereas SciVal uses publication information from the more comprehensive Scopus [12] database.

Bibliometrics exported from SciVal were chosen to provide evidence for three domains of bibliometrics: productivity, impact, and collaboration. A detailed list of metrics is presented in Table 1. Productivity metrics provide an overview of total scholarly output; in other words, how many publications a scholar produces within a time period. Impact metrics focus on citation count through raw and field-weighted or ratio values. Self-citations were excluded from analysis. Collaboration metrics consider all authors on a scholar’s publication with attention to co-authors’ affiliated institutions. These metrics represent both author-level and publication-level data. Author-level metrics, such as total number of publications, are those which directly quantify the output of the author, providing an overview of the author’s overall publication portfolio. Publication- or article-level metrics, such as citation rates, are representative of the performance of an individual publication. However, some article-level metrics when aggregated or normed, such as the number of citations for one publication, may show author-level information such as the author’s average number of citations which encompasses all publications. Information on the level of the metrics used in this study is denoted in Table 1.

Table 1. Summary of metrics used for analysis

* Represents a publication or article-level metric.

+ Represents an author-level metric – note many metrics can be article level but were analyzed in the aggregate or as normed-values of author-level data for this study.

** Self-citations were excluded from this analysis.

*** The most-cited journals are defined by the journal metrics CiteScore, Source-Normalized Impact per Paper, or SCImago Journal Rank.

RCR, relative citation ratio.

Three measures of grant funding were also collected: NIH funding as PI, Co-PI, Co-Investigator at the 5 year, 8 year, and overall time points. These measures of follow-on funding include only NIH federal funding received by a scholar identified as participating in the KL2, K08, or K23 program at CWRU from 2005 to 2013. Each of the three measures is a dichotomous variable of NIH federal funding received. Records of NIH funding can be found on NIH RePORTER [13].

Procedures

This study received institutional review board (IRB) exemption (IRB #20190435). The research team contacted the Office of Grants and Contracts for a list of all mentored-K award scholars at the institution since 2005. A librarian identified author IDs from Scopus/SciVal and provided SciVal training to the research team. Cohorts were created in SciVal based on the type of K award (KL2 or K08/K23) and the year the scholar began their K-grant (i.e., all 2005 KL2’s represent one cohort). All metrics from SciVal and iCite were extracted and entered into an SPSS file within a one month period. Grant data were collected using records of federal funding which were confirmed through NIH Reporter.

Analysis

Statistics were computed using IBM’s SPSS version 25 (IBM Corp., Armonk, NY, USA). Due to the variability of these data, there were several outliers for each metric. Non-parametric tests (Mann–Whitney U) were used when data were not normally distributed and medians (Mdn) and interquartile ranges (IQR) provided. Descriptive and inferential statistics were used to analyze the bibliometrics of KL2 scholars according to demographic groups and the bibliometric performance of KL2 scholars and K08/K23 scholars at the same institution. Federal funding data were analyzed using logistic rgression to assess the relationship between any significant variables found in the bibliometric analysis and a series of dichotomous funding variables.

Data were analyzed at three time points after matriculation: 5 years, 8 years [Reference Jagsi15,Reference Yin16], and overall (i.e., from the start year of a scholar’s K award through the end of 2018, regardless of the number of years). All data were time bound to include the year the scholar started their K award through the completion of 2018. iCite data were not included for the 5- and 8-year time points. This is because the main iCite metric, the relative citation ratio, is subject to latency, which means that the metric needs more time to develop and is not accurate for “new” publications [Reference Surkis and Spore17]. One KL2 scholar and one K08 were excluded from analyses due to lack of data. The 8-year matriculation analysis included only scholars who began their K award in 2011 or earlier.

Results

Eighty-five K Scholars met the criteria for analyses across the K08, K23, and KL2 grant mechanisms. K08 and K23 scholars were collapsed into one group for analysis to increase sample size equivalence. Additionally, combining the K08 and K23 scholars into one group more accurately matches the KL2 scholars’ research domains, which include behavioral, translational, and clinical research. Differences in the demographic composition of groups were analyzed. There was a significant difference between scholar degrees, with KL2 scholars holding more PhDs and the K08/K23 scholars more likely to have MDs or MD/PhDs (X2(12) = 27.59; p = .006). However, the majority of scholars in both groups held MDs. A summary of demographic information for KL2 and K08/K23 scholars is reported in Table 2.

Table 2. Summary of scholar demographics

* Indicates a cohort with a scholar excluded from analysis.

Note: One KL2 scholar from 2008 and one K08 scholar from 2012 were excluded from analysis due to lack of data. Bold text indicates the highest value.

Research Question 1: Are there demographic differences (gender, race, and degree) in bibliometric performance of KL2 scholars?

There were no significant differences in productivity, impact, or collaboration between male and female KL2 scholars, between Caucasian and Asian/African American KL2 scholars, or between scholars with clinical degrees (i.e., MDs) and those with clinical training and/or other professional/doctoral degrees (i.e., PhDs) at any time point. Ethnicity was not analyzed as the composition of the sample was 93% White, Non-Hispanic/Latino. No analyses between cohorts (e.g., the 2005 cohort and the 2006 cohort) were conducted due to the limited sample size per cohort (data not shown).

Research Question 2: Is there a difference between the bibliometric performance of KL2 scholars compared to other mentored-K awardees (K08 and K23) at the same institution?

Five years after matriculation, there were significant differences between KL2 and K08/K23 scholars on one productivity metric and one collaboration metric (Table 3). Examining scholarly output, KL2 scholars (Mdn = 15; IQR = 21.50) published significantly more than K08/K23 scholars (Mdn = 10; IQR = 10.50) during these 5 years (U = 608.50; p = .03). Measuring collaboration, KL2 scholars (M = .75; SD = .45) demonstrated higher field-weighted collaborations at the national level (t(81) = 2.12; p = .03) than K08/K23 scholars (M = .55; SD = .37).

Table 3. Results of 5-Year Metrics for KL2 and K08/K23 Scholars (N = 83)

*p < .05

Note: To indicate descriptive statistics, some notation was used: Mdn (median), M (mean), were reported for non-normal distributions. Bold text indicates the highest value.

An analysis of KL2 and K08/K23 scholars with 8 years of bibliometric data after matriculation again revealed significantly higher scholarly output (U = 412.50; p = .04) in KL2 scholars (Mdn = 27; IQR = 34.25) than in K08/K23 scholars (Mdn = 20; IQR = 20). In contrast, 8 years after matriculation, K08/K23 scholars (Mdn = 19.23; IQR = 19.59) had more international collaborations (U = 392.50; p = .024) than KL2 scholars (Mdn = 8.33; IQR = 22.39) (Table 4).

Table 4. Results of 8-Year Metrics for KL2 and K08/K23 Scholars (N = 63)

*p < .05

Note: To indicate descriptive statistics, some notation was used: Mdn (median), M (mean), were reported for non-normal distributions. Bold text indicates the highest value.

Analyses of the overall time point (i.e., all years from matriculation to the end of 2018) is reported in Table 5. In contrast to the 5- and 8-year metrics, overall (from the year a scholar started a K award through year end of 2018), there were no significant differences between KL2 and K08/K23 scholars in scholarly output. There was a significant difference between KL2 and K08/K23 scholars on two impact metrics. Overall, K08/K23 scholars (Mdn = 135.9; IQR = 26.33) had significantly more citations per publication (U = 609; p = .018) than KL2 scholars (Mdn = 268.20; IQR = 13.90), and K08/K23 scholars (M = 51.49; SD = 19.8) were published more in top percentile journals (t(81) = 2.255; p = .027) than KL2 scholars (M = 41.81; SD = 18.84). There was a significant difference between KL2 and K08/K23 scholars on two collaboration metrics. Overall, K08/K23 scholars (Mdn = 24.95; IQR = 21.57) had higher international collaboration (U = 618.50; p = .047) than KL2 scholars (Mdn = 12.80; IQR = 26.30), but KL2 scholars (M = .86; SD = .36) had a higher field-weighted national collaboration (t(81) = 2.122; p = .037) than K08/K23 scholars (M = .77; SD = .36).

Table 5. Results of Overall Metric for KL2 and K08/K23 Scholars (N = 83)

* p < .05

Note: To indicate descriptive statistics, some notation was used: Mdn (median), M (mean), were reported for non-normal distributions. Bold text indicates the highest value.

RCR, relative citation ratio.

Research Question 3: Is there a difference between post-K award federal grant funding (i.e., NIH) of KL2 scholars compared to other mentored-K awardees (K08 and K23) at the same institution?

Descriptive statistics indicated that 5 years after matriculation 32% of KL2 scholars and 31% of K08/K23 scholars had received NIH funding. NIH funding rates at the 8-year time point revealed that 48% of KL2s and 51% of K08/K23 scholars received NIH funding. To inform regression analyses, Pearson correlations were run using the significant variables found in the above inferential tests and grant funding outcomes. Several bibliometric variables were significantly positively correlated with NIH funding; however, the majority of the correlations were weak (<.03). Scholarly output (i.e., number of publications) was the strongest correlating variable across all time points for both KL2 and K08/K23 scholars (r = .3). Logistic Regression analyses reported that 5-year scholarly output significantly predicted KL2 funding outcomes (X2(1) = 5.26; p = .03; Nagelkerke R2 = 14%), but was non-significant for K08/K23 scholars (X2(1) = 2.61; p = .12). Overall total scholarly output (i.e., all years) reported no significant prediction for KL2 funding (X2(1) = 4.02; p = .06), but was significant for K08/K23 funding (X2(1) = 6.93; p = .03; Nagelkerke R2 = 24%).

Discussion

The results of this analysis suggest that the KL2 scholars have an advantage over K08 and K23 scholars at the 5-year time point in regards to NIH funding and scholarly output, as measured by bibliometric performance. This advantage is largely driven by the number of publications KL2s have and their field-weighted collaborations, when compared to the other K scholars at the same institution. Over time, these bibliometric measures shift and indicate little difference between the two groups, but with K08/K23s reporting more international collaborations and publications in top-tier journals.

It is important to focus on the 5-year time point and the factors that may have influence on the NIH funding rate. The relationship between publications and funding has been investigated for decades in relation to return on investment [Reference Lindsey18Reference Jacob and Lefgren20]. In other words, funders consider if the investigator will publish the results of their project and evidence for this is a strong publication record prior to funding. KL2 scholars are able to provide a “larger” publication history before their K08/K23 counterparts. This is consistent with some previous research which suggests that K awards who published nine publications per K year were significantly more likely to convert to R-funding compared to those with four publications [Reference Narahari5]; however, this research has thus far focused on K awardees who are surgeons. Additional research into the predictive relationship between publications and R-funding is needed for other early-career CTR investigators, such as KL2s.

This study’s findings indicating the initial high volume of KL2 publications may also be a reflection of the KL2 program’s structure and time for scholars to develop pilot data, which can be used in publications and grant applications. This means that immediately upon completion of the K-grant, the KL2s have a more established research presence (i.e., publications and pilot data), which initially leads to more funding. However, this initial advantage appears to diminish over time. Investigation into publication patterns and citation rates for clinicians versus non-clinician scientists should be explored. Further research into the factors that contribute to continued NIH funding for K-scholars should be considered.

The implications of using bibliometrics to evaluate translational research must also be considered by the research and evaluation community. In order to discuss this, we must first consider the definition of translational research. Understanding the difference between clinical and translational studies will help CSTAs and evaluators make informed decisions about how they evaluate scholarly performance when using publication data. The NIH defines clinical research as (1) patient-oriented with direct interaction with human subjects, (2) epidemiology and/or behavioral, or (3) outcomes or health services research [21]. The NIH definition of translational research is less precise and does not have clearly defined categories [22]; however, some translational science evaluators have interpreted the NIH definition of translational science to be: “a unidirectional continuum in which research findings are moved from the researcher’s bench to the patient’s bedside and community. In the continuum, the first stage of translational research (T1) transfers knowledge from basic research to clinical research, while the second stage (T2) transfers findings from clinical studies or clinical trials to practice settings and communities, where the findings improve health” [Reference Rubio23]. It is clear that elements of clinical (and basic) science are undoubtedly present in the definition of translational science, but the NIH’s definition of translational science alludes to a broader translational continuum. It is worth noting that some translational scientists and evaluators feel strongly that the NIH’s definition should also emphasis that translational research exists on a multidirectional continuum [Reference Rubio23].

Given the distinction between CTR, we must consider the validity of using the same bibliometrics to evaluate translational research in comparison to clinical research. The use of these metrics has been questioned over the last 15 years with many of the original pitfalls still existing today [Reference Wallin24]. One example of a pitfall, regarding a commonly reported metric used to evaluate both fields of research, is the percentage of publications in a top-tier journal. Other examples of such pitfalls include the varying length of time to publication by journal, the types of articles accepted, and the number of references and citations permitted, all of which are factors that influence bibliometric data. Some CTSA evaluators have begun to determine which bibliometrics most appropriately capture translational research in order to accurately evaluate its impact [Reference Llewellyn25,Reference Schneider26]. While these studies provide guidance on translational research evaluation, the need to evaluate research from the training perspective, at the scholar-level and program, still remains.

This study was limited in its generalizability as it only focused on K scholars affiliated with a single academic institution. However, in order to assess the performance of KL2 scholars, it was an initial way to evaluate the impact of a CTR training program using a comparison group as an internal control. It should be noted that while there some are similarities between the K08/K23 and KL2 scholar programs, there are numerous differences. The K08/K23 group represented the best control group for this research as the scholars received training at the same institution. Additional limitations on gathering and comparing more granular grant funding information also limited the scope of these results. At this time, only NIH funding was investigated, but future analyses need to consider a variety of grant funding.

Given the results of this study and the 5-year KL2 advantage reported, it is necessary to consider how bibliometrics are interpreted within translational science and the evaluation of CTR training. Over time, differences between translational and clinical bibliometrics must be interpreted with caution and appropriate metrics for translational science must be established. Research must continue to investigate the performance and outcomes of CTR training across the country in order to demonstrate the value of training on the broader scientific community. Future research will expand the current findings to include multi-institutional comparisons of bibliometric outcomes and follow-on funding with a larger and more varied sample of K scholars from different academic health centers.

Acknowledgments

This project was supported by the Clinical and Translational Science Collaborative of Cleveland which is funded by the National Institutes of Health (NIH), NCATS, Clinical and Translational Science Award (CTSA) grant, UL1TR002548. The content is solely the responsibility of the authors and do not necessarily represent the official views of the NIH. Thanks to Elizabeth (Liz) Bernal at the CWRU Kelvin Smith Library for her continued support, training, and knowledge. This work would not have been possible without her and the support of the library. Thanks to Nicole Llewellyn from Emory University’s Rollins School of Public Health for her invaluable feedback and time.

Disclosures

The authors have no conflicts of interest to declare.

References

Calvin-Naylor, NA, et al. Education and training of clinical and translational study investigators and research coordinators: a competency-based approach. Journal of Clinical and Translational Science 2017; 1: 1625.CrossRefGoogle Scholar
Sorkness, CA, et al. KL2 mentored career development programs at clinical and translational science award hubs: practices and outcomes. Journal of Clinical and Translational Science 2020; 4: 4352.CrossRefGoogle ScholarPubMed
Sweeney, C, et al. Transition to independence: characteristics and outcomes of mentored career development (KL2) scholars at clinical and translational science award institutions. Academic Medicine: Journal of the Association of American Medical Colleges 2017; 92: 556.CrossRefGoogle ScholarPubMed
Yin, HL, et al. Sustaining the clinical and translational research workforce: training and empowering the next generation of investigators. Academic Medicine 2015; 90: 861865.CrossRefGoogle ScholarPubMed
Narahari, AK, et al. Cardiothoracic and vascular surgeons achieve high rates of K award conversion into R01 funding. The Annals of Thoracic Surgery 2018; 106: 602607.CrossRefGoogle ScholarPubMed
Sayavedra, N, Hogle, JA, Moberg, DP. Using publication data to evaluate a Clinical and Translational Science Award (CTSA) career development program: early outcomes from KL2 scholars. Journal of Clinical and Translational Science 2017; 1: 352360.CrossRefGoogle ScholarPubMed
Llewellyn, N, et al. Scope, influence, and interdisciplinary collaboration: the publication portfolio of the NIH Clinical and Translational Science Awards (CTSA) Program from 2006 through 2017. Evaluation & the Health Professions. 2020; 43. doi: 10.1177/0163278719839435 CrossRefGoogle ScholarPubMed
Yu, F, et al. Bibliometrics approach to evaluating the research impact of CTSAs: a pilot study. Journal of Clinical and Translational Science 2017; 127. doi: 10.1017/cts.2020.29.Google Scholar
National Institutions of Health. Research Career Development Awards – Programs [Internet]. Office of Research Training and Career Development. (https://researchtraining.nih.gov/programs/career-development)Google Scholar
National Institutions of Health Office of Portfolio Analysis. iCite. (https://icite.od.nih.gov/analysis)Google Scholar
NIH RePORTER. National Institutes of Health, Office of Extramural Research. (https://www.report.nih.gov)Google Scholar
Hutchins, BI, et al. Relative Citation Ratio (RCR): a new metric that uses citation rates to measure influence at the article level. PLoS One Biology 2016; 6: 14.Google Scholar
Jagsi, R, et al. Sex differences in attainment of independent funding by career development awardees. Annals of Internal Medicine 2009; 151: 804811.CrossRefGoogle ScholarPubMed
Yin, HL, et al. Sustaining the clinical and translational research workforce: training and empowering the next generation of investigators. Academic Medicine 2015; 90: 861865.CrossRefGoogle ScholarPubMed
Surkis, A, Spore, S. The relative citation ratio: what is it and why should medical librarians care? Journal of the Medical Library Association 2018; 106: 508.CrossRefGoogle ScholarPubMed
Lindsey, D. The relationship between performance indicators for academic research and funding: developing a measure of return on investment in science. Scientometrics 1991; 20: 221234.CrossRefGoogle Scholar
Rigby, J. Systematic grant and funding body acknowledgement data for publications: new dimensions and new controversies for research policy and evaluation. Research Evaluation 2011; 20: 365375.CrossRefGoogle Scholar
Jacob, BA, Lefgren, L. The impact of research grant funding on scientific productivity. Journal of Public Economics 2011; 95: 11681177.CrossRefGoogle ScholarPubMed
Glossary of Terms for Human Subjects Protection and Inclusion Issues, based on the 1997 Report of the NIH Director’s Panel on Clinical Research, entry: “clinical research”. (http://grants.nih.gov/grants/peer/tree_glossary.pdf)Google Scholar
National Institutes of Health. Definitions under Subsection 1 (Research Objectives), Section I (Funding Opportunity Description), Part II (Full Text of Announcement), of RFA-RM-07-007: Institutional Clinical and Translational Science Award (U54) March 2007.Google Scholar
Rubio, DM, et al. Defining translational research: implications for training. Academic Medicine 2010; 85: 470.CrossRefGoogle ScholarPubMed
Wallin, JA. Bibliometric methods: pitfalls and possibilities. Basic & Clinical Pharmacology & Toxicology 2005; 97: 261275.CrossRefGoogle ScholarPubMed
Llewellyn, N, et al. Charting the publication and citation impact of the NIH clinical and translational science awards (CTSA) program from 2006 through 2016. Academic Medicine 2018; 93: 1162.Google Scholar
Schneider, M, et al. Feasibility of common bibliometrics in evaluating translational science. Journal of Clinical and Translational Science 2017; 1: 4552.CrossRefGoogle ScholarPubMed
Figure 0

Table 1. Summary of metrics used for analysis

Figure 1

Table 2. Summary of scholar demographics

Figure 2

Table 3. Results of 5-Year Metrics for KL2 and K08/K23 Scholars (N = 83)

Figure 3

Table 4. Results of 8-Year Metrics for KL2 and K08/K23 Scholars (N = 63)

Figure 4

Table 5. Results of Overall Metric for KL2 and K08/K23 Scholars (N = 83)