“…4 A pilot study indeed suggested that this Cartesian hypothesis may provide an appropriate methodology for developing prediction models of affect description. 2 A number of computational models are nowadays available that extract perception-related properties from musical audio such as onset (e.g., Klapuri, 1999;Smith, 1994), beat (e.g., Toiviainen, 2001;Large & Kolen, 1994;Scheirer, 1998;Laroche, 2003), consonance (e.g., Aures, 1985;Daniel & Weber, 1997;Leman, 2000a), pitch (e.g., Clarisse et al, 2002;De Mulder et al, 2004), harmony, tonality (e.g., Terhardt, 1974;Parncutt, 1989, Leman, 1995, 2000b, timbre (e.g., Cosi et al, 1994;Toiviainen, 1996;De Poli & Prandoni, 1997). 3 Apart from a preliminary study by Scheirer et al (2000) , we know of no other attempts that relate these and similar audioextracted structural features to affect-based description of music.…”