Recent technological developments have enabled us to synthesize images and sounds concurrently within single computers, even in real time, giving birth to novel and genuinely integrated audiovisual art forms (Hunt et al. 1998). But how should we organize and compose such works? Given certain soundscape, what would form an appropriate sequence of images to that soundscape? Given certain sequence of images, what soundscape is appropriate to it? If the image sequence and the soundscape are being created concurrently, how should we compose them? Authors have proposed different approaches to these questions (Whitney 1980; Hunt et al. 1998; Lokki et al. 1998; Rudi 1998; Kim and Lipscomb 2003; Gerhard and Hepting 2004; Yeo et al. 2004). These approaches differ significantly, and they are based on diverse principles, such as correspondence of aural to visual harmony, audiovisual modeling of mathematical principles, audiovisual rendering, data sonification, algorithmic control, and parameter space exploration. It is important to notice that there is no easy or correct solution, because the problem we must deal with lies in combining two entirely different media in time (Hunt et al. 1998). A fuzzy-logic approach to the challenge of composing both sound and moving image within coherent framework is proposed here as an alternative solution. This approach is based on fuzzy-logic model that enables flexible mapping of either aural or visual information onto the other, and it is able to generate complex audiovisual relationships by very simple means. This mapping strategy is inspired by two fundamental ideas: isomorphism and synaesthesia. Isomorphism applies when two complex structures can be mapped onto each other based on the fact that changes in one modality consistently cause changes in another modality (Hofstadter 1999). The word synaesthesia comes directly from the Greek syn (together) and aisthesis (perceive; Van Campen 1999), thus meaning a union of the senses. Synaesthesia occurs when stimulation in one sensory modality automatically triggers perception in second modality in the absence of any direct stimulation to this second modality (Harrison and Baron-Cohen 1997). This article is structured as follows. First, the motivations for this work are presented and discussed, including discussions of audiovisual domains, synaesthesia, and isomorphisms. Second, fuzzy logic is introduced, including its main features. Third, details of the proposed fuzzy-logic mapper are presented. Fourth, ID-FUSIONES (2001) and TIME EXPOSURE (2005), computer music-video works that use the proposed model, are discussed as actual implementations of the approach described in this article. Finally, conclusions and directions for future work are addressed.
Read full abstract