Automated generation of assessment tests from domain ontologies

The objective of this paper is to investigate the scope of OWL-DL ontologies in generating multiple choice questions (MCQs) that can be employed for conducting large scale assessments, and to conduct a detailed study on the effectiveness of the generated assessment items, using principles in the Ite...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Semantic Web 2017-01, Vol.8 (6), p.1023-1047
Hauptverfasser: E.V., Vinu, Kumar, P. Sreenivasa
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1047
container_issue 6
container_start_page 1023
container_title Semantic Web
container_volume 8
creator E.V., Vinu
Kumar, P. Sreenivasa
description The objective of this paper is to investigate the scope of OWL-DL ontologies in generating multiple choice questions (MCQs) that can be employed for conducting large scale assessments, and to conduct a detailed study on the effectiveness of the generated assessment items, using principles in the Item Response Theory (IRT). The details of a prototype system called Automatic Test Generation (ATG) system and its extended version called Extended-ATG system are elaborated. The ATG system (the initial system) was useful in generating multiple choice question-sets of required sizes from a given formal ontology. It works by employing a set of heuristics for selecting only those questions which are required for conducting a domain related assessment. We enhance this system with new features such as finding the difficulty values of generated MCQs and controlling the overall difficulty-level of question-sets, to form Extended-ATG system (the new system). This paper discusses the novel methods adopted to address these new features. That is, a method to determine the difficulty-level of a question-stem and an algorithm to control the difficulty of a question-set. While the ATG system uses at most two predicates for generating the stems of MCQs, the E-ATG system has no such limitations and employs several interesting predicate based patterns for stem generation. These predicate patterns are obtained from a detailed empirical study of large real-world question-sets. In addition, the new system also incorporates a specific non-pattern based approach which makes use of aggregation-like operations, to generate questions that involve superlatives (e.g., highest mountain, largest river etc.). We studied the feasibility and usefulness of the proposed methods by generating MCQs from several ontologies available online. The effectiveness of the suggested question selection heuristics is studied by comparing the resulting questions with those questions which were prepared by domain experts. It is found that the difficulty-scores of questions computed by the proposed system are highly correlated with their actual difficulty-scores determined with the help of IRT applied to data from classroom experiments. Our results show that the E-ATG system can generate domain specific question-sets which are close to the human generated ones (in terms of their semantic similarity). Also, the system can be potentially used for controlling the overall difficulty-level of the automatically genera
doi_str_mv 10.3233/SW-170252
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_1994005593</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1994005593</sourcerecordid><originalsourceid>FETCH-LOGICAL-c257t-29e7d70532ee1e17fba657005c14d7196c1a6bec034aa32d410530d43d2ecee3</originalsourceid><addsrcrecordid>eNotkEtrwzAQhEVpoSbNof_A0FMPbrV6WPaphNAXBHpIIEehSOvgEFupJB_676vizmUP8zE7DCH3QJ844_x5u69AUSbZFSkYA1qJtm6uSQFS0Yo2QtySZYwnmiWh5o0syMtqSn4wCV15xBGDSb0fS9-VJkaMccAxlQljimUX_FC6zPbZH5M_-2OP8Y7cdOYccfl_F2T39rpbf1Sbr_fP9WpTWSZVqliLyikqOUMEBNUdTJ1LUWlBOAVtbcHUB7SUC2M4cwIyS53gjqFF5AvyMMdegv-ech998lMY80cNbStykGx5ph5nygYfY8BOX0I_mPCjgeq_hfR2r-eF-C_wxlfB</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1994005593</pqid></control><display><type>article</type><title>Automated generation of assessment tests from domain ontologies</title><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>E.V., Vinu ; Kumar, P. Sreenivasa</creator><creatorcontrib>E.V., Vinu ; Kumar, P. Sreenivasa</creatorcontrib><description>The objective of this paper is to investigate the scope of OWL-DL ontologies in generating multiple choice questions (MCQs) that can be employed for conducting large scale assessments, and to conduct a detailed study on the effectiveness of the generated assessment items, using principles in the Item Response Theory (IRT). The details of a prototype system called Automatic Test Generation (ATG) system and its extended version called Extended-ATG system are elaborated. The ATG system (the initial system) was useful in generating multiple choice question-sets of required sizes from a given formal ontology. It works by employing a set of heuristics for selecting only those questions which are required for conducting a domain related assessment. We enhance this system with new features such as finding the difficulty values of generated MCQs and controlling the overall difficulty-level of question-sets, to form Extended-ATG system (the new system). This paper discusses the novel methods adopted to address these new features. That is, a method to determine the difficulty-level of a question-stem and an algorithm to control the difficulty of a question-set. While the ATG system uses at most two predicates for generating the stems of MCQs, the E-ATG system has no such limitations and employs several interesting predicate based patterns for stem generation. These predicate patterns are obtained from a detailed empirical study of large real-world question-sets. In addition, the new system also incorporates a specific non-pattern based approach which makes use of aggregation-like operations, to generate questions that involve superlatives (e.g., highest mountain, largest river etc.). We studied the feasibility and usefulness of the proposed methods by generating MCQs from several ontologies available online. The effectiveness of the suggested question selection heuristics is studied by comparing the resulting questions with those questions which were prepared by domain experts. It is found that the difficulty-scores of questions computed by the proposed system are highly correlated with their actual difficulty-scores determined with the help of IRT applied to data from classroom experiments. Our results show that the E-ATG system can generate domain specific question-sets which are close to the human generated ones (in terms of their semantic similarity). Also, the system can be potentially used for controlling the overall difficulty-level of the automatically generated question-sets for achieving specific pedagogical goals. However, our next challenge is to conduct a large-scale experiment under real-world conditions to study the psychometric characteristics (such as reliability and validity) of the automatically generated question items.</description><identifier>ISSN: 1570-0844</identifier><identifier>EISSN: 2210-4968</identifier><identifier>DOI: 10.3233/SW-170252</identifier><language>eng</language><publisher>Amsterdam: IOS Press BV</publisher><subject>Automation ; Item response theory ; Knowledge representation ; Ontology ; Programming languages ; Reliability aspects</subject><ispartof>Semantic Web, 2017-01, Vol.8 (6), p.1023-1047</ispartof><rights>Copyright IOS Press BV 2017</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c257t-29e7d70532ee1e17fba657005c14d7196c1a6bec034aa32d410530d43d2ecee3</citedby><cites>FETCH-LOGICAL-c257t-29e7d70532ee1e17fba657005c14d7196c1a6bec034aa32d410530d43d2ecee3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>E.V., Vinu</creatorcontrib><creatorcontrib>Kumar, P. Sreenivasa</creatorcontrib><title>Automated generation of assessment tests from domain ontologies</title><title>Semantic Web</title><description>The objective of this paper is to investigate the scope of OWL-DL ontologies in generating multiple choice questions (MCQs) that can be employed for conducting large scale assessments, and to conduct a detailed study on the effectiveness of the generated assessment items, using principles in the Item Response Theory (IRT). The details of a prototype system called Automatic Test Generation (ATG) system and its extended version called Extended-ATG system are elaborated. The ATG system (the initial system) was useful in generating multiple choice question-sets of required sizes from a given formal ontology. It works by employing a set of heuristics for selecting only those questions which are required for conducting a domain related assessment. We enhance this system with new features such as finding the difficulty values of generated MCQs and controlling the overall difficulty-level of question-sets, to form Extended-ATG system (the new system). This paper discusses the novel methods adopted to address these new features. That is, a method to determine the difficulty-level of a question-stem and an algorithm to control the difficulty of a question-set. While the ATG system uses at most two predicates for generating the stems of MCQs, the E-ATG system has no such limitations and employs several interesting predicate based patterns for stem generation. These predicate patterns are obtained from a detailed empirical study of large real-world question-sets. In addition, the new system also incorporates a specific non-pattern based approach which makes use of aggregation-like operations, to generate questions that involve superlatives (e.g., highest mountain, largest river etc.). We studied the feasibility and usefulness of the proposed methods by generating MCQs from several ontologies available online. The effectiveness of the suggested question selection heuristics is studied by comparing the resulting questions with those questions which were prepared by domain experts. It is found that the difficulty-scores of questions computed by the proposed system are highly correlated with their actual difficulty-scores determined with the help of IRT applied to data from classroom experiments. Our results show that the E-ATG system can generate domain specific question-sets which are close to the human generated ones (in terms of their semantic similarity). Also, the system can be potentially used for controlling the overall difficulty-level of the automatically generated question-sets for achieving specific pedagogical goals. However, our next challenge is to conduct a large-scale experiment under real-world conditions to study the psychometric characteristics (such as reliability and validity) of the automatically generated question items.</description><subject>Automation</subject><subject>Item response theory</subject><subject>Knowledge representation</subject><subject>Ontology</subject><subject>Programming languages</subject><subject>Reliability aspects</subject><issn>1570-0844</issn><issn>2210-4968</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><recordid>eNotkEtrwzAQhEVpoSbNof_A0FMPbrV6WPaphNAXBHpIIEehSOvgEFupJB_676vizmUP8zE7DCH3QJ844_x5u69AUSbZFSkYA1qJtm6uSQFS0Yo2QtySZYwnmiWh5o0syMtqSn4wCV15xBGDSb0fS9-VJkaMccAxlQljimUX_FC6zPbZH5M_-2OP8Y7cdOYccfl_F2T39rpbf1Sbr_fP9WpTWSZVqliLyikqOUMEBNUdTJ1LUWlBOAVtbcHUB7SUC2M4cwIyS53gjqFF5AvyMMdegv-ech998lMY80cNbStykGx5ph5nygYfY8BOX0I_mPCjgeq_hfR2r-eF-C_wxlfB</recordid><startdate>20170101</startdate><enddate>20170101</enddate><creator>E.V., Vinu</creator><creator>Kumar, P. Sreenivasa</creator><general>IOS Press BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20170101</creationdate><title>Automated generation of assessment tests from domain ontologies</title><author>E.V., Vinu ; Kumar, P. Sreenivasa</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c257t-29e7d70532ee1e17fba657005c14d7196c1a6bec034aa32d410530d43d2ecee3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Automation</topic><topic>Item response theory</topic><topic>Knowledge representation</topic><topic>Ontology</topic><topic>Programming languages</topic><topic>Reliability aspects</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>E.V., Vinu</creatorcontrib><creatorcontrib>Kumar, P. Sreenivasa</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Semantic Web</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>E.V., Vinu</au><au>Kumar, P. Sreenivasa</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Automated generation of assessment tests from domain ontologies</atitle><jtitle>Semantic Web</jtitle><date>2017-01-01</date><risdate>2017</risdate><volume>8</volume><issue>6</issue><spage>1023</spage><epage>1047</epage><pages>1023-1047</pages><issn>1570-0844</issn><eissn>2210-4968</eissn><abstract>The objective of this paper is to investigate the scope of OWL-DL ontologies in generating multiple choice questions (MCQs) that can be employed for conducting large scale assessments, and to conduct a detailed study on the effectiveness of the generated assessment items, using principles in the Item Response Theory (IRT). The details of a prototype system called Automatic Test Generation (ATG) system and its extended version called Extended-ATG system are elaborated. The ATG system (the initial system) was useful in generating multiple choice question-sets of required sizes from a given formal ontology. It works by employing a set of heuristics for selecting only those questions which are required for conducting a domain related assessment. We enhance this system with new features such as finding the difficulty values of generated MCQs and controlling the overall difficulty-level of question-sets, to form Extended-ATG system (the new system). This paper discusses the novel methods adopted to address these new features. That is, a method to determine the difficulty-level of a question-stem and an algorithm to control the difficulty of a question-set. While the ATG system uses at most two predicates for generating the stems of MCQs, the E-ATG system has no such limitations and employs several interesting predicate based patterns for stem generation. These predicate patterns are obtained from a detailed empirical study of large real-world question-sets. In addition, the new system also incorporates a specific non-pattern based approach which makes use of aggregation-like operations, to generate questions that involve superlatives (e.g., highest mountain, largest river etc.). We studied the feasibility and usefulness of the proposed methods by generating MCQs from several ontologies available online. The effectiveness of the suggested question selection heuristics is studied by comparing the resulting questions with those questions which were prepared by domain experts. It is found that the difficulty-scores of questions computed by the proposed system are highly correlated with their actual difficulty-scores determined with the help of IRT applied to data from classroom experiments. Our results show that the E-ATG system can generate domain specific question-sets which are close to the human generated ones (in terms of their semantic similarity). Also, the system can be potentially used for controlling the overall difficulty-level of the automatically generated question-sets for achieving specific pedagogical goals. However, our next challenge is to conduct a large-scale experiment under real-world conditions to study the psychometric characteristics (such as reliability and validity) of the automatically generated question items.</abstract><cop>Amsterdam</cop><pub>IOS Press BV</pub><doi>10.3233/SW-170252</doi><tpages>25</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1570-0844
ispartof Semantic Web, 2017-01, Vol.8 (6), p.1023-1047
issn 1570-0844
2210-4968
language eng
recordid cdi_proquest_journals_1994005593
source Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Automation
Item response theory
Knowledge representation
Ontology
Programming languages
Reliability aspects
title Automated generation of assessment tests from domain ontologies
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-11T01%3A25%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Automated%20generation%20of%20assessment%20tests%20from%20domain%20ontologies&rft.jtitle=Semantic%20Web&rft.au=E.V.,%20Vinu&rft.date=2017-01-01&rft.volume=8&rft.issue=6&rft.spage=1023&rft.epage=1047&rft.pages=1023-1047&rft.issn=1570-0844&rft.eissn=2210-4968&rft_id=info:doi/10.3233/SW-170252&rft_dat=%3Cproquest_cross%3E1994005593%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1994005593&rft_id=info:pmid/&rfr_iscdi=true