Show simple item record

dc.contributor.advisorWeinberg, Normanen
dc.contributor.authorBARSETTI-NERLAND, DANIEL EDWARD
dc.creatorBARSETTI-NERLAND, DANIEL EDWARDen
dc.date.accessioned2016-06-10T18:58:04Z
dc.date.available2016-06-10T18:58:04Z
dc.date.issued2016
dc.identifier.citationBARSETTI-NERLAND, DANIEL EDWARD. (2016). CREATING A RESPONSIVE VISUALIZATION THAT REACTS WITH MUSIC IN REAL TIME: INTEGRATING ABLETON LIVE 9 AND CYCLING ’74 MAX FOR LIVE INTO A MUSICAL PERFORMANCE (Bachelor's thesis, University of Arizona, Tucson, USA).
dc.identifier.urihttp://hdl.handle.net/10150/612546
dc.description.abstractOver the past three semesters, I have been working on refining my Honors Thesis to meet exactly what I wanted to study. This is why when looking at what I wanted to focus on it, it was obvious that I wanted to integrate musical performance with a visual performance—similar to what a Video Disc Jockey (VDJ) might incorporate into a live set. Currently, there are a quite a few percussion pieces that use Max patches that process the sounds, or use pre-made visuals that play along with the music. My goal with this Thesis was to find a way to use a patch with the music that reacts to the visuals in real time. What makes this different from other performances is that each performance will be slightly different than the one before it or after it. To me, this is something that I find very exciting about music and technology. Throughout this essay, I will explain how I approach composing a piece specifically for this as well as my learning process to integrate a Max patch with the Music.
dc.language.isoen_USen
dc.publisherThe University of Arizona.en
dc.rightsCopyright © is held by the author. Digital access to this material is made possible by the University Libraries, University of Arizona. Further transmission, reproduction or presentation (such as public display or performance) of protected items is prohibited except with permission of the author.en
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/
dc.titleCREATING A RESPONSIVE VISUALIZATION THAT REACTS WITH MUSIC IN REAL TIME: INTEGRATING ABLETON LIVE 9 AND CYCLING ’74 MAX FOR LIVE INTO A MUSICAL PERFORMANCEen_US
dc.typetexten
dc.typeElectronic Thesisen
thesis.degree.grantorUniversity of Arizonaen
thesis.degree.levelBachelorsen
thesis.degree.disciplineHonors Collegeen
thesis.degree.disciplineMusic Educationen
thesis.degree.nameB.M.en
refterms.dateFOA2018-09-11T12:26:31Z
html.description.abstractOver the past three semesters, I have been working on refining my Honors Thesis to meet exactly what I wanted to study. This is why when looking at what I wanted to focus on it, it was obvious that I wanted to integrate musical performance with a visual performance—similar to what a Video Disc Jockey (VDJ) might incorporate into a live set. Currently, there are a quite a few percussion pieces that use Max patches that process the sounds, or use pre-made visuals that play along with the music. My goal with this Thesis was to find a way to use a patch with the music that reacts to the visuals in real time. What makes this different from other performances is that each performance will be slightly different than the one before it or after it. To me, this is something that I find very exciting about music and technology. Throughout this essay, I will explain how I approach composing a piece specifically for this as well as my learning process to integrate a Max patch with the Music.


Files in this item

Thumbnail
Name:
azu_etd_mr_2016_0018_sip1_m.pdf
Size:
640.1Kb
Format:
PDF

This item appears in the following Collection(s)

Show simple item record