Functional Scaffolding for Composing Additional Musical Voices
Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users t...
Gespeichert in:
Veröffentlicht in: | Computer music journal 2014-12, Vol.38 (4), p.80-99 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 99 |
---|---|
container_issue | 4 |
container_start_page | 80 |
container_title | Computer music journal |
container_volume | 38 |
creator | Hoover, Amy K. Szerlip, Paul A. Stanley, Kenneth O. |
description | Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise. |
doi_str_mv | 10.1162/COMJ_a_00269 |
format | Article |
fullrecord | <record><control><sourceid>jstor_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1162_COMJ_a_00269</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>24265451</jstor_id><sourcerecordid>24265451</sourcerecordid><originalsourceid>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</originalsourceid><addsrcrecordid>eNptkEFv1DAQhS1EJZaWG1eklbhwIHTG9jjxBVGtWqBq1UOBq-U4DnKUrJc4QYJfX4et2hXq6Wnkb57fPMZeI3xAVPx0c3N9aawB4Eo_YyskgQWh0s_ZClBWheaqfMFeptQBgBKkVuzjxbx1U4hb269vnW3b2Ddh-3PdxnG9icMupmU6a5pwD13PKbisP2JwPp2wo9b2yb-612P2_eL82-ZLcXXz-evm7KpwksRUVCRROiJV-7qtyWosAbBuqGpdCQSN5gJL23qPCkvMIsBZBFc1KGty4pi92_vuxvhr9mkyQ0jO973d-jgng0rliwC1zOjb_9AuzmOOvlCSNEld6Uy931NujCmNvjW7MQx2_GMQzFKmOSwz4_LBtPNuGubkH31JSU6VuV0KX_pGKZeCy8fYQziM8fQPn55AXRy636IK0giQhJXhwDFvG9Dmb9j9ez-weLO36NIUx4eDuOSK8rK4A_rioWU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1645954989</pqid></control><display><type>article</type><title>Functional Scaffolding for Composing Additional Musical Voices</title><source>Jstor Complete Legacy</source><creator>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</creator><creatorcontrib>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</creatorcontrib><description>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</description><identifier>ISSN: 0148-9267</identifier><identifier>EISSN: 1531-5169</identifier><identifier>DOI: 10.1162/COMJ_a_00269</identifier><identifier>CODEN: CMUJDY</identifier><language>eng</language><publisher>One Rogers St., Cambridge, MA 02142-1209, USA: MIT Press</publisher><subject>Algorithmic Composition ; Composition ; Computer music ; Homo sapiens ; Human ; Laments ; Melody ; Music ; Music composition ; Musical pitch ; Musical rhythm ; Online media ; Polyphony ; Representations ; Rhythm ; Scaffolding ; Scaffolds ; Surface layer ; Texture ; Voice</subject><ispartof>Computer music journal, 2014-12, Vol.38 (4), p.80-99</ispartof><rights>2014 Massachusetts Institute of Technology</rights><rights>Copyright © Massachusetts Institute of Technology.</rights><rights>Copyright MIT Press Journals Winter 2014</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</citedby><cites>FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/24265451$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/24265451$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>314,776,780,799,27903,27904,57996,58229</link.rule.ids></links><search><creatorcontrib>Hoover, Amy K.</creatorcontrib><creatorcontrib>Szerlip, Paul A.</creatorcontrib><creatorcontrib>Stanley, Kenneth O.</creatorcontrib><title>Functional Scaffolding for Composing Additional Musical Voices</title><title>Computer music journal</title><description>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</description><subject>Algorithmic Composition</subject><subject>Composition</subject><subject>Computer music</subject><subject>Homo sapiens</subject><subject>Human</subject><subject>Laments</subject><subject>Melody</subject><subject>Music</subject><subject>Music composition</subject><subject>Musical pitch</subject><subject>Musical rhythm</subject><subject>Online media</subject><subject>Polyphony</subject><subject>Representations</subject><subject>Rhythm</subject><subject>Scaffolding</subject><subject>Scaffolds</subject><subject>Surface layer</subject><subject>Texture</subject><subject>Voice</subject><issn>0148-9267</issn><issn>1531-5169</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2014</creationdate><recordtype>article</recordtype><recordid>eNptkEFv1DAQhS1EJZaWG1eklbhwIHTG9jjxBVGtWqBq1UOBq-U4DnKUrJc4QYJfX4et2hXq6Wnkb57fPMZeI3xAVPx0c3N9aawB4Eo_YyskgQWh0s_ZClBWheaqfMFeptQBgBKkVuzjxbx1U4hb269vnW3b2Ddh-3PdxnG9icMupmU6a5pwD13PKbisP2JwPp2wo9b2yb-612P2_eL82-ZLcXXz-evm7KpwksRUVCRROiJV-7qtyWosAbBuqGpdCQSN5gJL23qPCkvMIsBZBFc1KGty4pi92_vuxvhr9mkyQ0jO973d-jgng0rliwC1zOjb_9AuzmOOvlCSNEld6Uy931NujCmNvjW7MQx2_GMQzFKmOSwz4_LBtPNuGubkH31JSU6VuV0KX_pGKZeCy8fYQziM8fQPn55AXRy636IK0giQhJXhwDFvG9Dmb9j9ez-weLO36NIUx4eDuOSK8rK4A_rioWU</recordid><startdate>20141201</startdate><enddate>20141201</enddate><creator>Hoover, Amy K.</creator><creator>Szerlip, Paul A.</creator><creator>Stanley, Kenneth O.</creator><general>MIT Press</general><general>The MIT Press</general><general>MIT Press Journals, The</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20141201</creationdate><title>Functional Scaffolding for Composing Additional Musical Voices</title><author>Hoover, Amy K. ; Szerlip, Paul A. ; Stanley, Kenneth O.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c453t-85414c556bebfb5a917001bd58fc7050d92317afee16171ee130ca10c8d14b5c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2014</creationdate><topic>Algorithmic Composition</topic><topic>Composition</topic><topic>Computer music</topic><topic>Homo sapiens</topic><topic>Human</topic><topic>Laments</topic><topic>Melody</topic><topic>Music</topic><topic>Music composition</topic><topic>Musical pitch</topic><topic>Musical rhythm</topic><topic>Online media</topic><topic>Polyphony</topic><topic>Representations</topic><topic>Rhythm</topic><topic>Scaffolding</topic><topic>Scaffolds</topic><topic>Surface layer</topic><topic>Texture</topic><topic>Voice</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hoover, Amy K.</creatorcontrib><creatorcontrib>Szerlip, Paul A.</creatorcontrib><creatorcontrib>Stanley, Kenneth O.</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computer music journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hoover, Amy K.</au><au>Szerlip, Paul A.</au><au>Stanley, Kenneth O.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Functional Scaffolding for Composing Additional Musical Voices</atitle><jtitle>Computer music journal</jtitle><date>2014-12-01</date><risdate>2014</risdate><volume>38</volume><issue>4</issue><spage>80</spage><epage>99</epage><pages>80-99</pages><issn>0148-9267</issn><eissn>1531-5169</eissn><coden>CMUJDY</coden><abstract>Many tools for computer-assisted composition contain built-in music-theoretical assumptions that may constrain the output to particular styles. In contrast, this article presents a new musical representation that contains almost no built-in knowledge, but that allows even musically untrained users to generate polyphonic textures that are derived from the user's own initial compositions. This representation, called functional scaffolding for musical composition (FSMC), exploits a simple yet powerful property of multipart compositions: The pattern of notes and rhythms in different instrumental parts of the same song are functionally related. That is, in principle, one part can be expressed as a function of another. Music in FSMC is represented accordingly as a functional relationship between an existing human composition, or scaffold, and a generated set of one or more additional musical voices. A human user without any musical expertise can then explore how the generated voice (or voices) should relate to the scaffold through an interactive evolutionary process akin to animal breeding. By inheriting from the intrinsic style and texture of the piece provided by the user, this approach can generate additional voices for potentially any style of music without the need for extensive musical expertise.</abstract><cop>One Rogers St., Cambridge, MA 02142-1209, USA</cop><pub>MIT Press</pub><doi>10.1162/COMJ_a_00269</doi><tpages>20</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0148-9267 |
ispartof | Computer music journal, 2014-12, Vol.38 (4), p.80-99 |
issn | 0148-9267 1531-5169 |
language | eng |
recordid | cdi_crossref_primary_10_1162_COMJ_a_00269 |
source | Jstor Complete Legacy |
subjects | Algorithmic Composition Composition Computer music Homo sapiens Human Laments Melody Music Music composition Musical pitch Musical rhythm Online media Polyphony Representations Rhythm Scaffolding Scaffolds Surface layer Texture Voice |
title | Functional Scaffolding for Composing Additional Musical Voices |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T13%3A28%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Functional%20Scaffolding%20for%20Composing%20Additional%20Musical%20Voices&rft.jtitle=Computer%20music%20journal&rft.au=Hoover,%20Amy%20K.&rft.date=2014-12-01&rft.volume=38&rft.issue=4&rft.spage=80&rft.epage=99&rft.pages=80-99&rft.issn=0148-9267&rft.eissn=1531-5169&rft.coden=CMUJDY&rft_id=info:doi/10.1162/COMJ_a_00269&rft_dat=%3Cjstor_cross%3E24265451%3C/jstor_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1645954989&rft_id=info:pmid/&rft_jstor_id=24265451&rfr_iscdi=true |