Crowd-driven Music: Interactive and Generative Approaches using Machine Vision and Manhattan

(FULL UP) Crowd-driven Music: Interactive and Generative Approaches using Machine Vision and Manhattan

19:00 – 21:00 (GMT+1)

Hosted by:

Chris Nash (Lead Organiser), UWE Bristol

Corey Ford (Technical Support), UWE Bristol

Maximum number of participants: 25

Abstract

This workshop offers a practical opportunity to code mappings from chaotic crowds to structured music using Manhattan (http://nash.audio/manhattan), an end-user programming environment and music editor designed to bridge composition and coding practices, coupled with live input from crowd footage, analysed using machine vision. Various techniques and examples (notably from BBC Music Day) are explored over two 1-hour sessions, respectively offering a practical introduction to crowd-driven music, followed by delegates’ development of original works. The workshop is accessible to anyone with a Mac/Windows computer, of any aesthetic and ability. An optional introduction to basic coding practices in Manhattan is additional offered.

Schedule:

  • (optional) Manhattan Primer (1-hour) *
  • Introduction to Crowd-driven Music (1-hour)
  • Crowd-driven Music Masterclass (1-hour)

URLs:

Crowd-driven Music Video Examples (mp4) – http://nash.audio/manhattan/nime2020/

Manhattan Software (for Mac and Windows) – http://nash.audio/manhattan

Bela.io link
Cycling 74 link