A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports

Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine lea...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-12
Hauptverfasser: Okasa, Gabriel, de León, Alberto, Strinzel, Michaela, Jorstad, Anne, Milzow, Katrin, Egger, Matthias, Müller, Stefan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Okasa, Gabriel
de León, Alberto
Strinzel, Michaela
Jorstad, Anne
Milzow, Katrin
Egger, Matthias
Müller, Stefan
description Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3133048613</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3133048613</sourcerecordid><originalsourceid>FETCH-proquest_journals_31330486133</originalsourceid><addsrcrecordid>eNqNjMsKwjAQRYMgWLT_EHBdSJO2dlvEx0JF1H0JOtUUSeJMW3_fCH6AqwPnHu6IRVKpNCkzKScsJmqFELJYyDxXETtU_Nx7wMEQ3PheXx_GAt-BRmvsnVfeowuSNw55RQREX71BbTt-BEB-gsHAO8A77GjGxo1-EsQ_Ttl8vbost0m4efVAXd26Hm2YapUqJbKyCPiv-gCQCj2x</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3133048613</pqid></control><display><type>article</type><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><source>Free E- Journals</source><creator>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</creator><creatorcontrib>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</creatorcontrib><description>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Annotations ; Categories ; Classification ; Human performance ; Machine learning ; Natural language processing ; Peer review ; Supervised learning ; Texts</subject><ispartof>arXiv.org, 2024-12</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Okasa, Gabriel</creatorcontrib><creatorcontrib>de León, Alberto</creatorcontrib><creatorcontrib>Strinzel, Michaela</creatorcontrib><creatorcontrib>Jorstad, Anne</creatorcontrib><creatorcontrib>Milzow, Katrin</creatorcontrib><creatorcontrib>Egger, Matthias</creatorcontrib><creatorcontrib>Müller, Stefan</creatorcontrib><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><title>arXiv.org</title><description>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</description><subject>Annotations</subject><subject>Categories</subject><subject>Classification</subject><subject>Human performance</subject><subject>Machine learning</subject><subject>Natural language processing</subject><subject>Peer review</subject><subject>Supervised learning</subject><subject>Texts</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNjMsKwjAQRYMgWLT_EHBdSJO2dlvEx0JF1H0JOtUUSeJMW3_fCH6AqwPnHu6IRVKpNCkzKScsJmqFELJYyDxXETtU_Nx7wMEQ3PheXx_GAt-BRmvsnVfeowuSNw55RQREX71BbTt-BEB-gsHAO8A77GjGxo1-EsQ_Ttl8vbost0m4efVAXd26Hm2YapUqJbKyCPiv-gCQCj2x</recordid><startdate>20241219</startdate><enddate>20241219</enddate><creator>Okasa, Gabriel</creator><creator>de León, Alberto</creator><creator>Strinzel, Michaela</creator><creator>Jorstad, Anne</creator><creator>Milzow, Katrin</creator><creator>Egger, Matthias</creator><creator>Müller, Stefan</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241219</creationdate><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><author>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31330486133</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Annotations</topic><topic>Categories</topic><topic>Classification</topic><topic>Human performance</topic><topic>Machine learning</topic><topic>Natural language processing</topic><topic>Peer review</topic><topic>Supervised learning</topic><topic>Texts</topic><toplevel>online_resources</toplevel><creatorcontrib>Okasa, Gabriel</creatorcontrib><creatorcontrib>de León, Alberto</creatorcontrib><creatorcontrib>Strinzel, Michaela</creatorcontrib><creatorcontrib>Jorstad, Anne</creatorcontrib><creatorcontrib>Milzow, Katrin</creatorcontrib><creatorcontrib>Egger, Matthias</creatorcontrib><creatorcontrib>Müller, Stefan</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Okasa, Gabriel</au><au>de León, Alberto</au><au>Strinzel, Michaela</au><au>Jorstad, Anne</au><au>Milzow, Katrin</au><au>Egger, Matthias</au><au>Müller, Stefan</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</atitle><jtitle>arXiv.org</jtitle><date>2024-12-19</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-12
issn 2331-8422
language eng
recordid cdi_proquest_journals_3133048613
source Free E- Journals
subjects Annotations
Categories
Classification
Human performance
Machine learning
Natural language processing
Peer review
Supervised learning
Texts
title A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T06%3A02%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20Supervised%20Machine%20Learning%20Approach%20for%20Assessing%20Grant%20Peer%20Review%20Reports&rft.jtitle=arXiv.org&rft.au=Okasa,%20Gabriel&rft.date=2024-12-19&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3133048613%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3133048613&rft_id=info:pmid/&rfr_iscdi=true