Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods

Issues surrounding the psychometric properties of writing assessments have received ongoing attention. However, the reliability estimates of scores derived from various holistic and analytical scoring strategies reported in the literature have relied on classical test theory (CT), which accounts for...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Educational and psychological measurement 1999-06, Vol.59 (3), p.492-506
Hauptverfasser: Swartz, Carl W., Hooper, Stephen R., Montgomery, James W., Wakely, Melissa B., de Kruif, Renee E. L., Reed, Martha, Brown, Timothy T., Levine, Melvin D., White, Kinnard P.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 506
container_issue 3
container_start_page 492
container_title Educational and psychological measurement
container_volume 59
creator Swartz, Carl W.
Hooper, Stephen R.
Montgomery, James W.
Wakely, Melissa B.
de Kruif, Renee E. L.
Reed, Martha
Brown, Timothy T.
Levine, Melvin D.
White, Kinnard P.
description Issues surrounding the psychometric properties of writing assessments have received ongoing attention. However, the reliability estimates of scores derived from various holistic and analytical scoring strategies reported in the literature have relied on classical test theory (CT), which accounts for only a single source of variance within a given analysis. Generalizability theory (GT) is a more powerful and flexible strategy that allows for the simultaneous estimation of multiple sources of error variance to estimate the reliability of test scores. Using GT, two studies were conducted to investigate the impact of the number of raters and the type of decision (relative vs. absolute) on the reliability of writing scores. The results of both studies indicated that the reliability coefficients for writing scores decline as (a) the number of raters is reduced and (b) when absolute decisions rather than relative decisions are made.
doi_str_mv 10.1177/00131649921970008
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_221559887</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ericid>EJ594286</ericid><sage_id>10.1177_00131649921970008</sage_id><sourcerecordid>41615778</sourcerecordid><originalsourceid>FETCH-LOGICAL-a408t-4e3f84a13bfc0fb54647ac2aa93868c096f2fafb5a9cbd32f5717402adc6f2883</originalsourceid><addsrcrecordid>eNp1kclKBDEURYMo2A4fILgI6rY0U1UlS9F2QhEccFm8TiV2pKxokhbalZ9u2nYCcRXCOTd53IfQBiW7lNb1HiGU00ooxaiqCSFyAQ1oWbKCSykX0WDGiyyIZbQS40M2iKB0gN5uo-vv8bHpTYDOvcLIdS5N8c3Y-DDFyeNhTO4RksFpbPCV6dyX4i2-Cy7N4tfaBxPxoQnuxbTYBv-IT3znclRj6Fu830M3zRfoPtxZ5sKksW_jGlqy0EWz_nmuotuj4c3BSXF-eXx6sH9egCAyFcJwKwVQPrKa2FEpKlGDZgCKy0pqoirLLGQASo9azmxZ01oQBq3OREq-irbm7z4F_zwxMTUPfhLyWLFhLDelpKyztP2fREUtGGeq5Nmic0sHH2MwtnkKuaIwbShpZtto_mwjZ3Y-X4aYW7ABeu3iT1AKURGWtc25lpvU33R4VirBZJXx7hxHuDe_Zvv323fiCqB0</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1474232953</pqid></control><display><type>article</type><title>Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods</title><source>SAGE Complete A-Z List</source><source>Periodicals Index Online</source><creator>Swartz, Carl W. ; Hooper, Stephen R. ; Montgomery, James W. ; Wakely, Melissa B. ; de Kruif, Renee E. L. ; Reed, Martha ; Brown, Timothy T. ; Levine, Melvin D. ; White, Kinnard P.</creator><creatorcontrib>Swartz, Carl W. ; Hooper, Stephen R. ; Montgomery, James W. ; Wakely, Melissa B. ; de Kruif, Renee E. L. ; Reed, Martha ; Brown, Timothy T. ; Levine, Melvin D. ; White, Kinnard P.</creatorcontrib><description>Issues surrounding the psychometric properties of writing assessments have received ongoing attention. However, the reliability estimates of scores derived from various holistic and analytical scoring strategies reported in the literature have relied on classical test theory (CT), which accounts for only a single source of variance within a given analysis. Generalizability theory (GT) is a more powerful and flexible strategy that allows for the simultaneous estimation of multiple sources of error variance to estimate the reliability of test scores. Using GT, two studies were conducted to investigate the impact of the number of raters and the type of decision (relative vs. absolute) on the reliability of writing scores. The results of both studies indicated that the reliability coefficients for writing scores decline as (a) the number of raters is reduced and (b) when absolute decisions rather than relative decisions are made.</description><identifier>ISSN: 0013-1644</identifier><identifier>EISSN: 1552-3888</identifier><identifier>DOI: 10.1177/00131649921970008</identifier><identifier>CODEN: EPMEAJ</identifier><language>eng</language><publisher>Thousand Oaks, CA: SAGE Publications</publisher><subject>Analytic Scoring ; Biological and medical sciences ; Educational evaluation ; Educational psychology ; Estimation (Mathematics) ; Fundamental and applied biological sciences. Psychology ; Generalizability Theory ; Holistic Evaluation ; Intermediate Grades ; Middle Schools ; Orientation. Evaluation ; Outcomes of Education ; Psychology. Psychoanalysis. Psychiatry ; Psychology. Psychophysiology ; Reliability ; Scores ; Scoring ; Test Theory ; Testing ; Tests ; Writing Tests</subject><ispartof>Educational and psychological measurement, 1999-06, Vol.59 (3), p.492-506</ispartof><rights>1999 INIST-CNRS</rights><rights>Copyright Sage Publications, Inc. Jun 1999</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-a408t-4e3f84a13bfc0fb54647ac2aa93868c096f2fafb5a9cbd32f5717402adc6f2883</citedby><cites>FETCH-LOGICAL-a408t-4e3f84a13bfc0fb54647ac2aa93868c096f2fafb5a9cbd32f5717402adc6f2883</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://journals.sagepub.com/doi/pdf/10.1177/00131649921970008$$EPDF$$P50$$Gsage$$H</linktopdf><linktohtml>$$Uhttps://journals.sagepub.com/doi/10.1177/00131649921970008$$EHTML$$P50$$Gsage$$H</linktohtml><link.rule.ids>314,776,780,21799,27848,27903,27904,43600,43601</link.rule.ids><backlink>$$Uhttp://eric.ed.gov/ERICWebPortal/detail?accno=EJ594286$$DView record in ERIC$$Hfree_for_read</backlink><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=1844602$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Swartz, Carl W.</creatorcontrib><creatorcontrib>Hooper, Stephen R.</creatorcontrib><creatorcontrib>Montgomery, James W.</creatorcontrib><creatorcontrib>Wakely, Melissa B.</creatorcontrib><creatorcontrib>de Kruif, Renee E. L.</creatorcontrib><creatorcontrib>Reed, Martha</creatorcontrib><creatorcontrib>Brown, Timothy T.</creatorcontrib><creatorcontrib>Levine, Melvin D.</creatorcontrib><creatorcontrib>White, Kinnard P.</creatorcontrib><title>Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods</title><title>Educational and psychological measurement</title><description>Issues surrounding the psychometric properties of writing assessments have received ongoing attention. However, the reliability estimates of scores derived from various holistic and analytical scoring strategies reported in the literature have relied on classical test theory (CT), which accounts for only a single source of variance within a given analysis. Generalizability theory (GT) is a more powerful and flexible strategy that allows for the simultaneous estimation of multiple sources of error variance to estimate the reliability of test scores. Using GT, two studies were conducted to investigate the impact of the number of raters and the type of decision (relative vs. absolute) on the reliability of writing scores. The results of both studies indicated that the reliability coefficients for writing scores decline as (a) the number of raters is reduced and (b) when absolute decisions rather than relative decisions are made.</description><subject>Analytic Scoring</subject><subject>Biological and medical sciences</subject><subject>Educational evaluation</subject><subject>Educational psychology</subject><subject>Estimation (Mathematics)</subject><subject>Fundamental and applied biological sciences. Psychology</subject><subject>Generalizability Theory</subject><subject>Holistic Evaluation</subject><subject>Intermediate Grades</subject><subject>Middle Schools</subject><subject>Orientation. Evaluation</subject><subject>Outcomes of Education</subject><subject>Psychology. Psychoanalysis. Psychiatry</subject><subject>Psychology. Psychophysiology</subject><subject>Reliability</subject><subject>Scores</subject><subject>Scoring</subject><subject>Test Theory</subject><subject>Testing</subject><subject>Tests</subject><subject>Writing Tests</subject><issn>0013-1644</issn><issn>1552-3888</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>1999</creationdate><recordtype>article</recordtype><sourceid>K30</sourceid><recordid>eNp1kclKBDEURYMo2A4fILgI6rY0U1UlS9F2QhEccFm8TiV2pKxokhbalZ9u2nYCcRXCOTd53IfQBiW7lNb1HiGU00ooxaiqCSFyAQ1oWbKCSykX0WDGiyyIZbQS40M2iKB0gN5uo-vv8bHpTYDOvcLIdS5N8c3Y-DDFyeNhTO4RksFpbPCV6dyX4i2-Cy7N4tfaBxPxoQnuxbTYBv-IT3znclRj6Fu830M3zRfoPtxZ5sKksW_jGlqy0EWz_nmuotuj4c3BSXF-eXx6sH9egCAyFcJwKwVQPrKa2FEpKlGDZgCKy0pqoirLLGQASo9azmxZ01oQBq3OREq-irbm7z4F_zwxMTUPfhLyWLFhLDelpKyztP2fREUtGGeq5Nmic0sHH2MwtnkKuaIwbShpZtto_mwjZ3Y-X4aYW7ABeu3iT1AKURGWtc25lpvU33R4VirBZJXx7hxHuDe_Zvv323fiCqB0</recordid><startdate>19990601</startdate><enddate>19990601</enddate><creator>Swartz, Carl W.</creator><creator>Hooper, Stephen R.</creator><creator>Montgomery, James W.</creator><creator>Wakely, Melissa B.</creator><creator>de Kruif, Renee E. L.</creator><creator>Reed, Martha</creator><creator>Brown, Timothy T.</creator><creator>Levine, Melvin D.</creator><creator>White, Kinnard P.</creator><general>SAGE Publications</general><general>Sage</general><general>Educational and Psychological Measurement, etc</general><general>SAGE PUBLICATIONS, INC</general><scope>7SW</scope><scope>BJH</scope><scope>BNH</scope><scope>BNI</scope><scope>BNJ</scope><scope>BNO</scope><scope>ERI</scope><scope>PET</scope><scope>REK</scope><scope>WWN</scope><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>EOLOZ</scope><scope>FKUCP</scope><scope>IOIBA</scope><scope>K30</scope><scope>PAAUG</scope><scope>PAWHS</scope><scope>PAWZZ</scope><scope>PAXOH</scope><scope>PBHAV</scope><scope>PBQSW</scope><scope>PBYQZ</scope><scope>PCIWU</scope><scope>PCMID</scope><scope>PCZJX</scope><scope>PDGRG</scope><scope>PDWWI</scope><scope>PETMR</scope><scope>PFVGT</scope><scope>PGXDX</scope><scope>PIHIL</scope><scope>PISVA</scope><scope>PJCTQ</scope><scope>PJTMS</scope><scope>PLCHJ</scope><scope>PMHAD</scope><scope>PNQDJ</scope><scope>POUND</scope><scope>PPLAD</scope><scope>PQAPC</scope><scope>PQCAN</scope><scope>PQCMW</scope><scope>PQEME</scope><scope>PQHKH</scope><scope>PQMID</scope><scope>PQNCT</scope><scope>PQNET</scope><scope>PQSCT</scope><scope>PQSET</scope><scope>PSVJG</scope><scope>PVMQY</scope><scope>PZGFC</scope></search><sort><creationdate>19990601</creationdate><title>Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods</title><author>Swartz, Carl W. ; Hooper, Stephen R. ; Montgomery, James W. ; Wakely, Melissa B. ; de Kruif, Renee E. L. ; Reed, Martha ; Brown, Timothy T. ; Levine, Melvin D. ; White, Kinnard P.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a408t-4e3f84a13bfc0fb54647ac2aa93868c096f2fafb5a9cbd32f5717402adc6f2883</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>1999</creationdate><topic>Analytic Scoring</topic><topic>Biological and medical sciences</topic><topic>Educational evaluation</topic><topic>Educational psychology</topic><topic>Estimation (Mathematics)</topic><topic>Fundamental and applied biological sciences. Psychology</topic><topic>Generalizability Theory</topic><topic>Holistic Evaluation</topic><topic>Intermediate Grades</topic><topic>Middle Schools</topic><topic>Orientation. Evaluation</topic><topic>Outcomes of Education</topic><topic>Psychology. Psychoanalysis. Psychiatry</topic><topic>Psychology. Psychophysiology</topic><topic>Reliability</topic><topic>Scores</topic><topic>Scoring</topic><topic>Test Theory</topic><topic>Testing</topic><topic>Tests</topic><topic>Writing Tests</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Swartz, Carl W.</creatorcontrib><creatorcontrib>Hooper, Stephen R.</creatorcontrib><creatorcontrib>Montgomery, James W.</creatorcontrib><creatorcontrib>Wakely, Melissa B.</creatorcontrib><creatorcontrib>de Kruif, Renee E. L.</creatorcontrib><creatorcontrib>Reed, Martha</creatorcontrib><creatorcontrib>Brown, Timothy T.</creatorcontrib><creatorcontrib>Levine, Melvin D.</creatorcontrib><creatorcontrib>White, Kinnard P.</creatorcontrib><collection>ERIC</collection><collection>ERIC (Ovid)</collection><collection>ERIC</collection><collection>ERIC</collection><collection>ERIC (Legacy Platform)</collection><collection>ERIC( SilverPlatter )</collection><collection>ERIC</collection><collection>ERIC PlusText (Legacy Platform)</collection><collection>Education Resources Information Center (ERIC)</collection><collection>ERIC</collection><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Periodicals Index Online Segment 01</collection><collection>Periodicals Index Online Segment 04</collection><collection>Periodicals Index Online Segment 29</collection><collection>Periodicals Index Online</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - West</collection><collection>Primary Sources Access (Plan D) - International</collection><collection>Primary Sources Access &amp; Build (Plan A) - MEA</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - Midwest</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - Northeast</collection><collection>Primary Sources Access (Plan D) - Southeast</collection><collection>Primary Sources Access (Plan D) - North Central</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - Southeast</collection><collection>Primary Sources Access (Plan D) - South Central</collection><collection>Primary Sources Access &amp; Build (Plan A) - UK / I</collection><collection>Primary Sources Access (Plan D) - Canada</collection><collection>Primary Sources Access (Plan D) - EMEALA</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - North Central</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - South Central</collection><collection>Primary Sources Access &amp; Build (Plan A) - International</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - International</collection><collection>Primary Sources Access (Plan D) - West</collection><collection>Periodicals Index Online Segments 1-50</collection><collection>Primary Sources Access (Plan D) - APAC</collection><collection>Primary Sources Access (Plan D) - Midwest</collection><collection>Primary Sources Access (Plan D) - MEA</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - Canada</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - UK / I</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - EMEALA</collection><collection>Primary Sources Access &amp; Build (Plan A) - APAC</collection><collection>Primary Sources Access &amp; Build (Plan A) - Canada</collection><collection>Primary Sources Access &amp; Build (Plan A) - West</collection><collection>Primary Sources Access &amp; Build (Plan A) - EMEALA</collection><collection>Primary Sources Access (Plan D) - Northeast</collection><collection>Primary Sources Access &amp; Build (Plan A) - Midwest</collection><collection>Primary Sources Access &amp; Build (Plan A) - North Central</collection><collection>Primary Sources Access &amp; Build (Plan A) - Northeast</collection><collection>Primary Sources Access &amp; Build (Plan A) - South Central</collection><collection>Primary Sources Access &amp; Build (Plan A) - Southeast</collection><collection>Primary Sources Access (Plan D) - UK / I</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - APAC</collection><collection>Primary Sources Access—Foundation Edition (Plan E) - MEA</collection><jtitle>Educational and psychological measurement</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Swartz, Carl W.</au><au>Hooper, Stephen R.</au><au>Montgomery, James W.</au><au>Wakely, Melissa B.</au><au>de Kruif, Renee E. L.</au><au>Reed, Martha</au><au>Brown, Timothy T.</au><au>Levine, Melvin D.</au><au>White, Kinnard P.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><ericid>EJ594286</ericid><atitle>Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods</atitle><jtitle>Educational and psychological measurement</jtitle><date>1999-06-01</date><risdate>1999</risdate><volume>59</volume><issue>3</issue><spage>492</spage><epage>506</epage><pages>492-506</pages><issn>0013-1644</issn><eissn>1552-3888</eissn><coden>EPMEAJ</coden><abstract>Issues surrounding the psychometric properties of writing assessments have received ongoing attention. However, the reliability estimates of scores derived from various holistic and analytical scoring strategies reported in the literature have relied on classical test theory (CT), which accounts for only a single source of variance within a given analysis. Generalizability theory (GT) is a more powerful and flexible strategy that allows for the simultaneous estimation of multiple sources of error variance to estimate the reliability of test scores. Using GT, two studies were conducted to investigate the impact of the number of raters and the type of decision (relative vs. absolute) on the reliability of writing scores. The results of both studies indicated that the reliability coefficients for writing scores decline as (a) the number of raters is reduced and (b) when absolute decisions rather than relative decisions are made.</abstract><cop>Thousand Oaks, CA</cop><pub>SAGE Publications</pub><doi>10.1177/00131649921970008</doi><tpages>15</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0013-1644
ispartof Educational and psychological measurement, 1999-06, Vol.59 (3), p.492-506
issn 0013-1644
1552-3888
language eng
recordid cdi_proquest_journals_221559887
source SAGE Complete A-Z List; Periodicals Index Online
subjects Analytic Scoring
Biological and medical sciences
Educational evaluation
Educational psychology
Estimation (Mathematics)
Fundamental and applied biological sciences. Psychology
Generalizability Theory
Holistic Evaluation
Intermediate Grades
Middle Schools
Orientation. Evaluation
Outcomes of Education
Psychology. Psychoanalysis. Psychiatry
Psychology. Psychophysiology
Reliability
Scores
Scoring
Test Theory
Testing
Tests
Writing Tests
title Using Generalizability Theory to Estimate the Reliability of Writing Scores Derived from Holistic and Analytical Scoring Methods
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T07%3A21%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Using%20Generalizability%20Theory%20to%20Estimate%20the%20Reliability%20of%20Writing%20Scores%20Derived%20from%20Holistic%20and%20Analytical%20Scoring%20Methods&rft.jtitle=Educational%20and%20psychological%20measurement&rft.au=Swartz,%20Carl%20W.&rft.date=1999-06-01&rft.volume=59&rft.issue=3&rft.spage=492&rft.epage=506&rft.pages=492-506&rft.issn=0013-1644&rft.eissn=1552-3888&rft.coden=EPMEAJ&rft_id=info:doi/10.1177/00131649921970008&rft_dat=%3Cproquest_cross%3E41615778%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1474232953&rft_id=info:pmid/&rft_ericid=EJ594286&rft_sage_id=10.1177_00131649921970008&rfr_iscdi=true