A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports
Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine lea...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-12 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Okasa, Gabriel de León, Alberto Strinzel, Michaela Jorstad, Anne Milzow, Katrin Egger, Matthias Müller, Stefan |
description | Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3133048613</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3133048613</sourcerecordid><originalsourceid>FETCH-proquest_journals_31330486133</originalsourceid><addsrcrecordid>eNqNjMsKwjAQRYMgWLT_EHBdSJO2dlvEx0JF1H0JOtUUSeJMW3_fCH6AqwPnHu6IRVKpNCkzKScsJmqFELJYyDxXETtU_Nx7wMEQ3PheXx_GAt-BRmvsnVfeowuSNw55RQREX71BbTt-BEB-gsHAO8A77GjGxo1-EsQ_Ttl8vbost0m4efVAXd26Hm2YapUqJbKyCPiv-gCQCj2x</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3133048613</pqid></control><display><type>article</type><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><source>Free E- Journals</source><creator>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</creator><creatorcontrib>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</creatorcontrib><description>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Annotations ; Categories ; Classification ; Human performance ; Machine learning ; Natural language processing ; Peer review ; Supervised learning ; Texts</subject><ispartof>arXiv.org, 2024-12</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Okasa, Gabriel</creatorcontrib><creatorcontrib>de León, Alberto</creatorcontrib><creatorcontrib>Strinzel, Michaela</creatorcontrib><creatorcontrib>Jorstad, Anne</creatorcontrib><creatorcontrib>Milzow, Katrin</creatorcontrib><creatorcontrib>Egger, Matthias</creatorcontrib><creatorcontrib>Müller, Stefan</creatorcontrib><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><title>arXiv.org</title><description>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</description><subject>Annotations</subject><subject>Categories</subject><subject>Classification</subject><subject>Human performance</subject><subject>Machine learning</subject><subject>Natural language processing</subject><subject>Peer review</subject><subject>Supervised learning</subject><subject>Texts</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNjMsKwjAQRYMgWLT_EHBdSJO2dlvEx0JF1H0JOtUUSeJMW3_fCH6AqwPnHu6IRVKpNCkzKScsJmqFELJYyDxXETtU_Nx7wMEQ3PheXx_GAt-BRmvsnVfeowuSNw55RQREX71BbTt-BEB-gsHAO8A77GjGxo1-EsQ_Ttl8vbost0m4efVAXd26Hm2YapUqJbKyCPiv-gCQCj2x</recordid><startdate>20241219</startdate><enddate>20241219</enddate><creator>Okasa, Gabriel</creator><creator>de León, Alberto</creator><creator>Strinzel, Michaela</creator><creator>Jorstad, Anne</creator><creator>Milzow, Katrin</creator><creator>Egger, Matthias</creator><creator>Müller, Stefan</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241219</creationdate><title>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</title><author>Okasa, Gabriel ; de León, Alberto ; Strinzel, Michaela ; Jorstad, Anne ; Milzow, Katrin ; Egger, Matthias ; Müller, Stefan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31330486133</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Annotations</topic><topic>Categories</topic><topic>Classification</topic><topic>Human performance</topic><topic>Machine learning</topic><topic>Natural language processing</topic><topic>Peer review</topic><topic>Supervised learning</topic><topic>Texts</topic><toplevel>online_resources</toplevel><creatorcontrib>Okasa, Gabriel</creatorcontrib><creatorcontrib>de León, Alberto</creatorcontrib><creatorcontrib>Strinzel, Michaela</creatorcontrib><creatorcontrib>Jorstad, Anne</creatorcontrib><creatorcontrib>Milzow, Katrin</creatorcontrib><creatorcontrib>Egger, Matthias</creatorcontrib><creatorcontrib>Müller, Stefan</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Okasa, Gabriel</au><au>de León, Alberto</au><au>Strinzel, Michaela</au><au>Jorstad, Anne</au><au>Milzow, Katrin</au><au>Egger, Matthias</au><au>Müller, Stefan</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports</atitle><jtitle>arXiv.org</jtitle><date>2024-12-19</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Peer review in grant evaluation informs funding decisions, but the contents of peer review reports are rarely analyzed. In this work, we develop a thoroughly tested pipeline to analyze the texts of grant peer review reports using methods from applied Natural Language Processing (NLP) and machine learning. We start by developing twelve categories reflecting content of grant peer review reports that are of interest to research funders. This is followed by multiple human annotators' iterative annotation of these categories in a novel text corpus of grant peer review reports submitted to the Swiss National Science Foundation. After validating the human annotation, we use the annotated texts to fine-tune pre-trained transformer models to classify these categories at scale, while conducting several robustness and validation checks. Our results show that many categories can be reliably identified by human annotators and machine learning approaches. However, the choice of text classification approach considerably influences the classification performance. We also find a high correspondence between out-of-sample classification performance and human annotators' perceived difficulty in identifying categories. Our results and publicly available fine-tuned transformer models will allow researchers and research funders and anybody interested in peer review to examine and report on the contents of these reports in a structured manner. Ultimately, we hope our approach can contribute to ensuring the quality and trustworthiness of grant peer review.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-12 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_3133048613 |
source | Free E- Journals |
subjects | Annotations Categories Classification Human performance Machine learning Natural language processing Peer review Supervised learning Texts |
title | A Supervised Machine Learning Approach for Assessing Grant Peer Review Reports |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T06%3A02%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20Supervised%20Machine%20Learning%20Approach%20for%20Assessing%20Grant%20Peer%20Review%20Reports&rft.jtitle=arXiv.org&rft.au=Okasa,%20Gabriel&rft.date=2024-12-19&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3133048613%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3133048613&rft_id=info:pmid/&rfr_iscdi=true |