Functional Scaffolding for Composing Additional Musical Voices

Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computer music journal 2014-12, Vol.38 (4), p.80-99
Hauptverfasser: Hoover, Amy K., Szerlip, Paul A., Stanley, Kenneth O.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 99
container_issue 4
container_start_page 80
container_title Computer music journal
container_volume 38
creator Hoover, Amy K.
Szerlip, Paul A.
Stanley, Kenneth O.
description Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.
doi_str_mv 10.1162/COMJ_a_00269
format Article
fullrecord <record><control><sourceid>jstor_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1162_COMJ_a_00269</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>24265451</jstor_id><sourcerecordid>24265451</sourcerecordid><originalsourceid>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</originalsourceid><addsrcrecordid>eNptkEFv1DAQhS1EJZaWG1eklbhwIHTG9jjxBVGtWqBq1UOBq-U4DnKUrJc4QYJfX4et2hXq6Wnkb57fPMZeI3xAVPx0c3N9aawB4Eo_YyskgQWh0s_ZClBWheaqfMFeptQBgBKkVuzjxbx1U4hb269vnW3b2Ddh-3PdxnG9icMupmU6a5pwD13PKbisP2JwPp2wo9b2yb-612P2_eL82-ZLcXXz-evm7KpwksRUVCRROiJV-7qtyWosAbBuqGpdCQSN5gJL23qPCkvMIsBZBFc1KGty4pi92_vuxvhr9mkyQ0jO973d-jgng0rliwC1zOjb_9AuzmOOvlCSNEld6Uy931NujCmNvjW7MQx2_GMQzFKmOSwz4_LBtPNuGubkH31JSU6VuV0KX_pGKZeCy8fYQziM8fQPn55AXRy636IK0giQhJXhwDFvG9Dmb9j9ez-weLO36NIUx4eDuOSK8rK4A_rioWU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1645954989</pqid></control><display><type>article</type><title>Functional Scaffolding for Composing Additional Musical Voices</title><source>Jstor Complete Legacy</source><creator>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</creator><creatorcontrib>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</creatorcontrib><description>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</description><identifier>ISSN: 0148-9267</identifier><identifier>EISSN: 1531-5169</identifier><identifier>DOI: 10.1162/COMJ_a_00269</identifier><identifier>CODEN: CMUJDY</identifier><language>eng</language><publisher>One Rogers St., Cambridge, MA 02142-1209, USA: MIT Press</publisher><subject>Algorithmic Composition ; Composition ; Computer music ; Homo sapiens ; Human ; Laments ; Melody ; Music ; Music composition ; Musical pitch ; Musical rhythm ; Online media ; Polyphony ; Representations ; Rhythm ; Scaffolding ; Scaffolds ; Surface layer ; Texture ; Voice</subject><ispartof>Computer music journal, 2014-12, Vol.38 (4), p.80-99</ispartof><rights>2014 Massachusetts Institute of Technology</rights><rights>Copyright © Massachusetts Institute of Technology.</rights><rights>Copyright MIT Press Journals Winter 2014</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</citedby><cites>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/24265451$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/24265451$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>314,776,780,799,27903,27904,57996,58229</link.rule.ids></links><search><creatorcontrib>Hoover, Amy K.</creatorcontrib><creatorcontrib>Szerlip, Paul A.</creatorcontrib><creatorcontrib>Stanley, Kenneth O.</creatorcontrib><title>Functional Scaffolding for Composing Additional Musical Voices</title><title>Computer music journal</title><description>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</description><subject>Algorithmic Composition</subject><subject>Composition</subject><subject>Computer music</subject><subject>Homo sapiens</subject><subject>Human</subject><subject>Laments</subject><subject>Melody</subject><subject>Music</subject><subject>Music composition</subject><subject>Musical pitch</subject><subject>Musical rhythm</subject><subject>Online media</subject><subject>Polyphony</subject><subject>Representations</subject><subject>Rhythm</subject><subject>Scaffolding</subject><subject>Scaffolds</subject><subject>Surface layer</subject><subject>Texture</subject><subject>Voice</subject><issn>0148-9267</issn><issn>1531-5169</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2014</creationdate><recordtype>article</recordtype><recordid>eNptkEFv1DAQhS1EJZaWG1eklbhwIHTG9jjxBVGtWqBq1UOBq-U4DnKUrJc4QYJfX4et2hXq6Wnkb57fPMZeI3xAVPx0c3N9aawB4Eo_YyskgQWh0s_ZClBWheaqfMFeptQBgBKkVuzjxbx1U4hb269vnW3b2Ddh-3PdxnG9icMupmU6a5pwD13PKbisP2JwPp2wo9b2yb-612P2_eL82-ZLcXXz-evm7KpwksRUVCRROiJV-7qtyWosAbBuqGpdCQSN5gJL23qPCkvMIsBZBFc1KGty4pi92_vuxvhr9mkyQ0jO973d-jgng0rliwC1zOjb_9AuzmOOvlCSNEld6Uy931NujCmNvjW7MQx2_GMQzFKmOSwz4_LBtPNuGubkH31JSU6VuV0KX_pGKZeCy8fYQziM8fQPn55AXRy636IK0giQhJXhwDFvG9Dmb9j9ez-weLO36NIUx4eDuOSK8rK4A_rioWU</recordid><startdate>20141201</startdate><enddate>20141201</enddate><creator>Hoover, Amy K.</creator><creator>Szerlip, Paul A.</creator><creator>Stanley, Kenneth O.</creator><general>MIT Press</general><general>The MIT Press</general><general>MIT Press Journals, The</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20141201</creationdate><title>Functional Scaffolding for Composing Additional Musical Voices</title><author>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2014</creationdate><topic>Algorithmic Composition</topic><topic>Composition</topic><topic>Computer music</topic><topic>Homo sapiens</topic><topic>Human</topic><topic>Laments</topic><topic>Melody</topic><topic>Music</topic><topic>Music composition</topic><topic>Musical pitch</topic><topic>Musical rhythm</topic><topic>Online media</topic><topic>Polyphony</topic><topic>Representations</topic><topic>Rhythm</topic><topic>Scaffolding</topic><topic>Scaffolds</topic><topic>Surface layer</topic><topic>Texture</topic><topic>Voice</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hoover, Amy K.</creatorcontrib><creatorcontrib>Szerlip, Paul A.</creatorcontrib><creatorcontrib>Stanley, Kenneth O.</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computer music journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hoover, Amy K.</au><au>Szerlip, Paul A.</au><au>Stanley, Kenneth O.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Functional Scaffolding for Composing Additional Musical Voices</atitle><jtitle>Computer music journal</jtitle><date>2014-12-01</date><risdate>2014</risdate><volume>38</volume><issue>4</issue><spage>80</spage><epage>99</epage><pages>80-99</pages><issn>0148-9267</issn><eissn>1531-5169</eissn><coden>CMUJDY</coden><abstract>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</abstract><cop>One Rogers St., Cambridge, MA 02142-1209, USA</cop><pub>MIT Press</pub><doi>10.1162/COMJ_a_00269</doi><tpages>20</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0148-9267
ispartof Computer music journal, 2014-12, Vol.38 (4), p.80-99
issn 0148-9267
1531-5169
language eng
recordid cdi_crossref_primary_10_1162_COMJ_a_00269
source Jstor Complete Legacy
subjects Algorithmic Composition
Composition
Computer music
Homo sapiens
Human
Laments
Melody
Music
Music composition
Musical pitch
Musical rhythm
Online media
Polyphony
Representations
Rhythm
Scaffolding
Scaffolds
Surface layer
Texture
Voice
title Functional Scaffolding for Composing Additional Musical Voices
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T13%3A28%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Functional%20Scaffolding%20for%20Composing%20Additional%20Musical%20Voices&rft.jtitle=Computer%20music%20journal&rft.au=Hoover,%20Amy%20K.&rft.date=2014-12-01&rft.volume=38&rft.issue=4&rft.spage=80&rft.epage=99&rft.pages=80-99&rft.issn=0148-9267&rft.eissn=1531-5169&rft.coden=CMUJDY&rft_id=info:doi/10.1162/COMJ_a_00269&rft_dat=%3Cjstor_cross%3E24265451%3C/jstor_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1645954989&rft_id=info:pmid/&rft_jstor_id=24265451&rfr_iscdi=true