Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams
Background When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students...
Gespeichert in:
Veröffentlicht in: | Journal of computer assisted learning 2024-12, Vol.40 (6), p.2667-2680 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 2680 |
---|---|
container_issue | 6 |
container_start_page | 2667 |
container_title | Journal of computer assisted learning |
container_volume | 40 |
creator | Pijeira‐Díaz, Héctor J. Subramanya, Shashank Pol, Janneke Bruin, Anique |
description | Background
When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability.
Objectives
In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain.
Methods
For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments.
Results and Conclusions
In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached.
Implications
Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload.
Lay Description
What is currently known about this topic?
Students' monitoring accuracy of causal relation learning is on average low.
Completing causal diagrams improves monitoring accuracy to some extent.
Advances in natural language processing (NLP) enable automated diagram assessment.
NLP‐powered learning analytics can be used for automated formative feedback.
What does this paper add?
Evaluation of the reliability of the automated diagram assessment.
Performance comparison of different language technologies and techniques.
The accuracy of the automated diagram assessment ranged from 84% to 86%.
Human‐computer Cohen's κ surpassed that |
doi_str_mv | 10.1111/jcal.12992 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3128087165</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3128087165</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</originalsourceid><addsrcrecordid>eNp9kE1OwzAQhS0EEqWw4QSWWCAhpdhOY8dLqMqfKiFBWVsTx6lS5ad4EqruOAJn5CS4hDWzebP45unNI-ScswkPc722UE240FockBGPZRIJJfQhGTEhZTTVTB-TE8Q1Y0xpmY5IPf-AqoeubFb01TWda6z7_vy6nb8sg2zarfMup5UD3-wRaKDadaVFWrSeQt-1NXQBAESHWAcD2hYUuz4PK15SCz1CRfMSVh5qPCVHBVTozv50TN7u5svZQ7R4vn-c3SwiK7QUURbnieUqixUXfCqnGcuVC9l1lqSOpyIVCdcysUw4nUuutC6USLSNrcgKmbF4TC4G341v33uHnVm3vQ_Z0cRcpCxVXCaBuhoo61tE7wqz8WUNfmc4M_s6zb5O81tngPkAb8vK7f4hzVN4ZLj5AfqXegM</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3128087165</pqid></control><display><type>article</type><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><source>Wiley Online Library Journals Frontfile Complete</source><creator>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</creator><creatorcontrib>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</creatorcontrib><description>Background
When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability.
Objectives
In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain.
Methods
For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments.
Results and Conclusions
In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached.
Implications
Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload.
Lay Description
What is currently known about this topic?
Students' monitoring accuracy of causal relation learning is on average low.
Completing causal diagrams improves monitoring accuracy to some extent.
Advances in natural language processing (NLP) enable automated diagram assessment.
NLP‐powered learning analytics can be used for automated formative feedback.
What does this paper add?
Evaluation of the reliability of the automated diagram assessment.
Performance comparison of different language technologies and techniques.
The accuracy of the automated diagram assessment ranged from 84% to 86%.
Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84).
Implications for practice/or policy
The tested technologies can be embedded into digital learning environments.
Diagram assessment can be reliably (semi‐)automated.
This can reduce teachers' workload and enable real‐time formative feedback.
This evaluation enables testing feedback interventions in future work.</description><identifier>ISSN: 0266-4909</identifier><identifier>EISSN: 1365-2729</identifier><identifier>DOI: 10.1111/jcal.12992</identifier><language>eng</language><publisher>Chichester, UK: John Wiley & Sons, Inc</publisher><subject>Accuracy ; automated formative feedback ; Automation ; causal diagrams ; Educational Assessment ; Feedback ; Feedback (Response) ; Formative evaluation ; Influence of Technology ; Language Processing ; Learning analytics ; Machine learning ; Monitoring ; Natural language processing ; Network reliability ; Neural networks ; Performance evaluation ; Position indicators ; sentence BERT ; Student Evaluation ; Students ; Support vector machines ; Teachers ; Workload ; Workloads</subject><ispartof>Journal of computer assisted learning, 2024-12, Vol.40 (6), p.2667-2680</ispartof><rights>2024 The Authors. published by John Wiley & Sons Ltd.</rights><rights>2024. This article is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</cites><orcidid>0000-0001-5178-0287 ; 0009-0005-2485-681X ; 0000-0003-2275-6397 ; 0000-0003-4580-8997</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://onlinelibrary.wiley.com/doi/pdf/10.1111%2Fjcal.12992$$EPDF$$P50$$Gwiley$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://onlinelibrary.wiley.com/doi/full/10.1111%2Fjcal.12992$$EHTML$$P50$$Gwiley$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,1411,27903,27904,45553,45554</link.rule.ids></links><search><creatorcontrib>Pijeira‐Díaz, Héctor J.</creatorcontrib><creatorcontrib>Subramanya, Shashank</creatorcontrib><creatorcontrib>Pol, Janneke</creatorcontrib><creatorcontrib>Bruin, Anique</creatorcontrib><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><title>Journal of computer assisted learning</title><description>Background
When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability.
Objectives
In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain.
Methods
For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments.
Results and Conclusions
In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached.
Implications
Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload.
Lay Description
What is currently known about this topic?
Students' monitoring accuracy of causal relation learning is on average low.
Completing causal diagrams improves monitoring accuracy to some extent.
Advances in natural language processing (NLP) enable automated diagram assessment.
NLP‐powered learning analytics can be used for automated formative feedback.
What does this paper add?
Evaluation of the reliability of the automated diagram assessment.
Performance comparison of different language technologies and techniques.
The accuracy of the automated diagram assessment ranged from 84% to 86%.
Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84).
Implications for practice/or policy
The tested technologies can be embedded into digital learning environments.
Diagram assessment can be reliably (semi‐)automated.
This can reduce teachers' workload and enable real‐time formative feedback.
This evaluation enables testing feedback interventions in future work.</description><subject>Accuracy</subject><subject>automated formative feedback</subject><subject>Automation</subject><subject>causal diagrams</subject><subject>Educational Assessment</subject><subject>Feedback</subject><subject>Feedback (Response)</subject><subject>Formative evaluation</subject><subject>Influence of Technology</subject><subject>Language Processing</subject><subject>Learning analytics</subject><subject>Machine learning</subject><subject>Monitoring</subject><subject>Natural language processing</subject><subject>Network reliability</subject><subject>Neural networks</subject><subject>Performance evaluation</subject><subject>Position indicators</subject><subject>sentence BERT</subject><subject>Student Evaluation</subject><subject>Students</subject><subject>Support vector machines</subject><subject>Teachers</subject><subject>Workload</subject><subject>Workloads</subject><issn>0266-4909</issn><issn>1365-2729</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>24P</sourceid><sourceid>WIN</sourceid><recordid>eNp9kE1OwzAQhS0EEqWw4QSWWCAhpdhOY8dLqMqfKiFBWVsTx6lS5ad4EqruOAJn5CS4hDWzebP45unNI-ScswkPc722UE240FockBGPZRIJJfQhGTEhZTTVTB-TE8Q1Y0xpmY5IPf-AqoeubFb01TWda6z7_vy6nb8sg2zarfMup5UD3-wRaKDadaVFWrSeQt-1NXQBAESHWAcD2hYUuz4PK15SCz1CRfMSVh5qPCVHBVTozv50TN7u5svZQ7R4vn-c3SwiK7QUURbnieUqixUXfCqnGcuVC9l1lqSOpyIVCdcysUw4nUuutC6USLSNrcgKmbF4TC4G341v33uHnVm3vQ_Z0cRcpCxVXCaBuhoo61tE7wqz8WUNfmc4M_s6zb5O81tngPkAb8vK7f4hzVN4ZLj5AfqXegM</recordid><startdate>202412</startdate><enddate>202412</enddate><creator>Pijeira‐Díaz, Héctor J.</creator><creator>Subramanya, Shashank</creator><creator>Pol, Janneke</creator><creator>Bruin, Anique</creator><general>John Wiley & Sons, Inc</general><general>Wiley Subscription Services, Inc</general><scope>24P</scope><scope>WIN</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-5178-0287</orcidid><orcidid>https://orcid.org/0009-0005-2485-681X</orcidid><orcidid>https://orcid.org/0000-0003-2275-6397</orcidid><orcidid>https://orcid.org/0000-0003-4580-8997</orcidid></search><sort><creationdate>202412</creationdate><title>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</title><author>Pijeira‐Díaz, Héctor J. ; Subramanya, Shashank ; Pol, Janneke ; Bruin, Anique</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2962-b3d5c17b37121464b0d7e9099b58e1828251965c02e9d61799f7259c3c2bf6b03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>automated formative feedback</topic><topic>Automation</topic><topic>causal diagrams</topic><topic>Educational Assessment</topic><topic>Feedback</topic><topic>Feedback (Response)</topic><topic>Formative evaluation</topic><topic>Influence of Technology</topic><topic>Language Processing</topic><topic>Learning analytics</topic><topic>Machine learning</topic><topic>Monitoring</topic><topic>Natural language processing</topic><topic>Network reliability</topic><topic>Neural networks</topic><topic>Performance evaluation</topic><topic>Position indicators</topic><topic>sentence BERT</topic><topic>Student Evaluation</topic><topic>Students</topic><topic>Support vector machines</topic><topic>Teachers</topic><topic>Workload</topic><topic>Workloads</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Pijeira‐Díaz, Héctor J.</creatorcontrib><creatorcontrib>Subramanya, Shashank</creatorcontrib><creatorcontrib>Pol, Janneke</creatorcontrib><creatorcontrib>Bruin, Anique</creatorcontrib><collection>Wiley Online Library Open Access</collection><collection>Wiley Free Content</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Journal of computer assisted learning</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Pijeira‐Díaz, Héctor J.</au><au>Subramanya, Shashank</au><au>Pol, Janneke</au><au>Bruin, Anique</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams</atitle><jtitle>Journal of computer assisted learning</jtitle><date>2024-12</date><risdate>2024</risdate><volume>40</volume><issue>6</issue><spage>2667</spage><epage>2680</epage><pages>2667-2680</pages><issn>0266-4909</issn><eissn>1365-2729</eissn><abstract>Background
When learning causal relations, completing causal diagrams enhances students' comprehension judgements to some extent. To potentially boost this effect, advances in natural language processing (NLP) enable real‐time formative feedback based on the automated assessment of students' diagrams, which can involve the correctness of both the responses and their position in the causal chain. However, the responsible adoption and effectiveness of automated diagram assessment depend on its reliability.
Objectives
In this study, we compare two Dutch pre‐trained models (i.e., based on RobBERT and BERTje) in combination with two machine‐learning classifiers—Support Vector Machine (SVM) and Neural Networks (NN), in terms of different indicators of automated diagram assessment reliability. We also contrast two techniques (i.e., semantic similarity and machine learning) for estimating the correct position of a student diagram response in the causal chain.
Methods
For training and evaluation of the models, we capitalize on a human‐labelled dataset containing 2900+ causal diagrams completed by 700+ secondary school students, accumulated from previous diagramming experiments.
Results and Conclusions
In predicting correct responses, 86% accuracy and Cohen's κ of 0.69 were reached, with combinations using SVM being roughly three‐times faster (important for real‐time applications) than their NN counterparts. In terms of predicting the response position in the causal diagrams, 92% accuracy and 0.89 Cohen's κ were reached.
Implications
Taken together, these evaluation figures equip educational designers for decision‐making on when these NLP‐powered learning analytics are warranted for automated formative feedback in causal relation learning; thereby potentially enabling real‐time feedback for learners and reducing teachers' workload.
Lay Description
What is currently known about this topic?
Students' monitoring accuracy of causal relation learning is on average low.
Completing causal diagrams improves monitoring accuracy to some extent.
Advances in natural language processing (NLP) enable automated diagram assessment.
NLP‐powered learning analytics can be used for automated formative feedback.
What does this paper add?
Evaluation of the reliability of the automated diagram assessment.
Performance comparison of different language technologies and techniques.
The accuracy of the automated diagram assessment ranged from 84% to 86%.
Human‐computer Cohen's κ surpassed that of human–human (0.89 vs. 0.84).
Implications for practice/or policy
The tested technologies can be embedded into digital learning environments.
Diagram assessment can be reliably (semi‐)automated.
This can reduce teachers' workload and enable real‐time formative feedback.
This evaluation enables testing feedback interventions in future work.</abstract><cop>Chichester, UK</cop><pub>John Wiley & Sons, Inc</pub><doi>10.1111/jcal.12992</doi><tpages>14</tpages><orcidid>https://orcid.org/0000-0001-5178-0287</orcidid><orcidid>https://orcid.org/0009-0005-2485-681X</orcidid><orcidid>https://orcid.org/0000-0003-2275-6397</orcidid><orcidid>https://orcid.org/0000-0003-4580-8997</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0266-4909 |
ispartof | Journal of computer assisted learning, 2024-12, Vol.40 (6), p.2667-2680 |
issn | 0266-4909 1365-2729 |
language | eng |
recordid | cdi_proquest_journals_3128087165 |
source | Wiley Online Library Journals Frontfile Complete |
subjects | Accuracy automated formative feedback Automation causal diagrams Educational Assessment Feedback Feedback (Response) Formative evaluation Influence of Technology Language Processing Learning analytics Machine learning Monitoring Natural language processing Network reliability Neural networks Performance evaluation Position indicators sentence BERT Student Evaluation Students Support vector machines Teachers Workload Workloads |
title | Evaluating Sentence‐BERT‐powered learning analytics for automated assessment of students' causal diagrams |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-21T12%3A40%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluating%20Sentence%E2%80%90BERT%E2%80%90powered%20learning%20analytics%20for%20automated%20assessment%20of%20students'%20causal%20diagrams&rft.jtitle=Journal%20of%20computer%20assisted%20learning&rft.au=Pijeira%E2%80%90D%C3%ADaz,%20H%C3%A9ctor%20J.&rft.date=2024-12&rft.volume=40&rft.issue=6&rft.spage=2667&rft.epage=2680&rft.pages=2667-2680&rft.issn=0266-4909&rft.eissn=1365-2729&rft_id=info:doi/10.1111/jcal.12992&rft_dat=%3Cproquest_cross%3E3128087165%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3128087165&rft_id=info:pmid/&rfr_iscdi=true |