Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams

Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of computer assisted learning 2024-12, Vol.40 (6), p.2667-2680
Hauptverfasser: Pijeira‐Díaz, Héctor J., Subramanya, Shashank, Pol, Janneke, Bruin, Anique
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 2680
container_issue 6
container_start_page 2667
container_title Journal of computer assisted learning
container_volume 40
creator Pijeira‐Díaz, Héctor J.
Subramanya, Shashank
Pol, Janneke
Bruin, Anique
description Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability. Objectives In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain. Methods For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments. Results and Conclusions In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached. Implications Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload. Lay Description What is currently known about this topic? Students' monitoring accuracy of causal relation learning is on average low. Completing causal diagrams improves monitoring accuracy to some extent. Advances in natural language processing (NLP) enable automated diagram assessment. NLP‐powered learning analytics can be used for automated formative feedback. What does this paper add? Evaluation of the reliability of the automated diagram assessment. Performance comparison of different language technologies and techniques. The accuracy of the automated diagram assessment ranged from 84% to 86%. Human‐computer Cohen's κ surpassed that
doi_str_mv 10.1111/jcal.12992
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3128087165</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3128087165</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</originalsourceid><addsrcrecordid>eNp9kE1OwzAQhS0EEqWw4QSWWCAhpdhOY8dLqMqfKiFBWVsTx6lS5ad4EqruOAJn5CS4hDWzebP45unNI-ScswkPc722UE240FockBGPZRIJJfQhGTEhZTTVTB-TE8Q1Y0xpmY5IPf-AqoeubFb01TWda6z7_vy6nb8sg2zarfMup5UD3-wRaKDadaVFWrSeQt-1NXQBAESHWAcD2hYUuz4PK15SCz1CRfMSVh5qPCVHBVTozv50TN7u5svZQ7R4vn-c3SwiK7QUURbnieUqixUXfCqnGcuVC9l1lqSOpyIVCdcysUw4nUuutC6USLSNrcgKmbF4TC4G341v33uHnVm3vQ_Z0cRcpCxVXCaBuhoo61tE7wqz8WUNfmc4M_s6zb5O81tngPkAb8vK7f4hzVN4ZLj5AfqXegM</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3128087165</pqid></control><display><type>article</type><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><source>Wiley Online Library Journals Frontfile Complete</source><creator>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</creator><creatorcontrib>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</creatorcontrib><description>Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability. Objectives In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain. Methods For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments. Results and Conclusions In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached. Implications Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload. Lay Description What is currently known about this topic? Students' monitoring accuracy of causal relation learning is on average low. Completing causal diagrams improves monitoring accuracy to some extent. Advances in natural language processing (NLP) enable automated diagram assessment. NLP‐powered learning analytics can be used for automated formative feedback. What does this paper add? Evaluation of the reliability of the automated diagram assessment. Performance comparison of different language technologies and techniques. The accuracy of the automated diagram assessment ranged from 84% to 86%. Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84). Implications for practice/or policy The tested technologies can be embedded into digital learning environments. Diagram assessment can be reliably (semi‐)automated. This can reduce teachers' workload and enable real‐time formative feedback. This evaluation enables testing feedback interventions in future work.</description><identifier>ISSN: 0266-4909</identifier><identifier>EISSN: 1365-2729</identifier><identifier>DOI: 10.1111/jcal.12992</identifier><language>eng</language><publisher>Chichester, UK: John Wiley &amp; Sons, Inc</publisher><subject>Accuracy ; automated formative feedback ; Automation ; causal diagrams ; Educational Assessment ; Feedback ; Feedback (Response) ; Formative evaluation ; Influence of Technology ; Language Processing ; Learning analytics ; Machine learning ; Monitoring ; Natural language processing ; Network reliability ; Neural networks ; Performance evaluation ; Position indicators ; sentence BERT ; Student Evaluation ; Students ; Support vector machines ; Teachers ; Workload ; Workloads</subject><ispartof>Journal of computer assisted learning, 2024-12, Vol.40 (6), p.2667-2680</ispartof><rights>2024 The Authors. published by John Wiley &amp; Sons Ltd.</rights><rights>2024. This article is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</cites><orcidid>0000-0001-5178-0287 ; 0009-0005-2485-681X ; 0000-0003-2275-6397 ; 0000-0003-4580-8997</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://onlinelibrary.wiley.com/doi/pdf/10.1111%2Fjcal.12992$$EPDF$$P50$$Gwiley$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://onlinelibrary.wiley.com/doi/full/10.1111%2Fjcal.12992$$EHTML$$P50$$Gwiley$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,1411,27903,27904,45553,45554</link.rule.ids></links><search><creatorcontrib>Pijeira‐Díaz, Héctor J.</creatorcontrib><creatorcontrib>Subramanya, Shashank</creatorcontrib><creatorcontrib>Pol, Janneke</creatorcontrib><creatorcontrib>Bruin, Anique</creatorcontrib><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><title>Journal of computer assisted learning</title><description>Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability. Objectives In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain. Methods For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments. Results and Conclusions In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached. Implications Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload. Lay Description What is currently known about this topic? Students' monitoring accuracy of causal relation learning is on average low. Completing causal diagrams improves monitoring accuracy to some extent. Advances in natural language processing (NLP) enable automated diagram assessment. NLP‐powered learning analytics can be used for automated formative feedback. What does this paper add? Evaluation of the reliability of the automated diagram assessment. Performance comparison of different language technologies and techniques. The accuracy of the automated diagram assessment ranged from 84% to 86%. Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84). Implications for practice/or policy The tested technologies can be embedded into digital learning environments. Diagram assessment can be reliably (semi‐)automated. This can reduce teachers' workload and enable real‐time formative feedback. This evaluation enables testing feedback interventions in future work.</description><subject>Accuracy</subject><subject>automated formative feedback</subject><subject>Automation</subject><subject>causal diagrams</subject><subject>Educational Assessment</subject><subject>Feedback</subject><subject>Feedback (Response)</subject><subject>Formative evaluation</subject><subject>Influence of Technology</subject><subject>Language Processing</subject><subject>Learning analytics</subject><subject>Machine learning</subject><subject>Monitoring</subject><subject>Natural language processing</subject><subject>Network reliability</subject><subject>Neural networks</subject><subject>Performance evaluation</subject><subject>Position indicators</subject><subject>sentence BERT</subject><subject>Student Evaluation</subject><subject>Students</subject><subject>Support vector machines</subject><subject>Teachers</subject><subject>Workload</subject><subject>Workloads</subject><issn>0266-4909</issn><issn>1365-2729</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>24P</sourceid><sourceid>WIN</sourceid><recordid>eNp9kE1OwzAQhS0EEqWw4QSWWCAhpdhOY8dLqMqfKiFBWVsTx6lS5ad4EqruOAJn5CS4hDWzebP45unNI-ScswkPc722UE240FockBGPZRIJJfQhGTEhZTTVTB-TE8Q1Y0xpmY5IPf-AqoeubFb01TWda6z7_vy6nb8sg2zarfMup5UD3-wRaKDadaVFWrSeQt-1NXQBAESHWAcD2hYUuz4PK15SCz1CRfMSVh5qPCVHBVTozv50TN7u5svZQ7R4vn-c3SwiK7QUURbnieUqixUXfCqnGcuVC9l1lqSOpyIVCdcysUw4nUuutC6USLSNrcgKmbF4TC4G341v33uHnVm3vQ_Z0cRcpCxVXCaBuhoo61tE7wqz8WUNfmc4M_s6zb5O81tngPkAb8vK7f4hzVN4ZLj5AfqXegM</recordid><startdate>202412</startdate><enddate>202412</enddate><creator>Pijeira‐Díaz, Héctor J.</creator><creator>Subramanya, Shashank</creator><creator>Pol, Janneke</creator><creator>Bruin, Anique</creator><general>John Wiley &amp; Sons, Inc</general><general>Wiley Subscription Services, Inc</general><scope>24P</scope><scope>WIN</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-5178-0287</orcidid><orcidid>https://orcid.org/0009-0005-2485-681X</orcidid><orcidid>https://orcid.org/0000-0003-2275-6397</orcidid><orcidid>https://orcid.org/0000-0003-4580-8997</orcidid></search><sort><creationdate>202412</creationdate><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><author>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>automated formative feedback</topic><topic>Automation</topic><topic>causal diagrams</topic><topic>Educational Assessment</topic><topic>Feedback</topic><topic>Feedback (Response)</topic><topic>Formative evaluation</topic><topic>Influence of Technology</topic><topic>Language Processing</topic><topic>Learning analytics</topic><topic>Machine learning</topic><topic>Monitoring</topic><topic>Natural language processing</topic><topic>Network reliability</topic><topic>Neural networks</topic><topic>Performance evaluation</topic><topic>Position indicators</topic><topic>sentence BERT</topic><topic>Student Evaluation</topic><topic>Students</topic><topic>Support vector machines</topic><topic>Teachers</topic><topic>Workload</topic><topic>Workloads</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Pijeira‐Díaz, Héctor J.</creatorcontrib><creatorcontrib>Subramanya, Shashank</creatorcontrib><creatorcontrib>Pol, Janneke</creatorcontrib><creatorcontrib>Bruin, Anique</creatorcontrib><collection>Wiley Online Library Open Access</collection><collection>Wiley Free Content</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Journal of computer assisted learning</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Pijeira‐Díaz, Héctor J.</au><au>Subramanya, Shashank</au><au>Pol, Janneke</au><au>Bruin, Anique</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</atitle><jtitle>Journal of computer assisted learning</jtitle><date>2024-12</date><risdate>2024</risdate><volume>40</volume><issue>6</issue><spage>2667</spage><epage>2680</epage><pages>2667-2680</pages><issn>0266-4909</issn><eissn>1365-2729</eissn><abstract>Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability. Objectives In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain. Methods For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments. Results and Conclusions In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached. Implications Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload. Lay Description What is currently known about this topic? Students' monitoring accuracy of causal relation learning is on average low. Completing causal diagrams improves monitoring accuracy to some extent. Advances in natural language processing (NLP) enable automated diagram assessment. NLP‐powered learning analytics can be used for automated formative feedback. What does this paper add? Evaluation of the reliability of the automated diagram assessment. Performance comparison of different language technologies and techniques. The accuracy of the automated diagram assessment ranged from 84% to 86%. Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84). Implications for practice/or policy The tested technologies can be embedded into digital learning environments. Diagram assessment can be reliably (semi‐)automated. This can reduce teachers' workload and enable real‐time formative feedback. This evaluation enables testing feedback interventions in future work.</abstract><cop>Chichester, UK</cop><pub>John Wiley &amp; Sons, Inc</pub><doi>10.1111/jcal.12992</doi><tpages>14</tpages><orcidid>https://orcid.org/0000-0001-5178-0287</orcidid><orcidid>https://orcid.org/0009-0005-2485-681X</orcidid><orcidid>https://orcid.org/0000-0003-2275-6397</orcidid><orcidid>https://orcid.org/0000-0003-4580-8997</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0266-4909
ispartof Journal of computer assisted learning, 2024-12, Vol.40 (6), p.2667-2680
issn 0266-4909
1365-2729
language eng
recordid cdi_proquest_journals_3128087165
source Wiley Online Library Journals Frontfile Complete
subjects Accuracy
automated formative feedback
Automation
causal diagrams
Educational Assessment
Feedback
Feedback (Response)
Formative evaluation
Influence of Technology
Language Processing
Learning analytics
Machine learning
Monitoring
Natural language processing
Network reliability
Neural networks
Performance evaluation
Position indicators
sentence BERT
Student Evaluation
Students
Support vector machines
Teachers
Workload
Workloads
title Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-21T12%3A40%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluating%20Sentence%E2%80%90BERT%E2%80%90powered%20learning%20analytics%20for%20automated%20assessment%20of%20students'%20causal%20diagrams&rft.jtitle=Journal%20of%20computer%20assisted%20learning&rft.au=Pijeira%E2%80%90D%C3%ADaz,%20H%C3%A9ctor%20J.&rft.date=2024-12&rft.volume=40&rft.issue=6&rft.spage=2667&rft.epage=2680&rft.pages=2667-2680&rft.issn=0266-4909&rft.eissn=1365-2729&rft_id=info:doi/10.1111/jcal.12992&rft_dat=%3Cproquest_cross%3E3128087165%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3128087165&rft_id=info:pmid/&rfr_iscdi=true