AISB Convention 2015

The AISB Convention is an annual conference covering the range of AI and Cognitive Science, organised by the Society for the Study of Artificial Intelligence and Simulation of Behaviour. The 2015 Convention will be held at the Uni...


Yasemin Erden on BBC

AISB Committee member, and Philosophy Programme Director and Lecturer, Dr Yasemin J. Erden interviewed for the BBC on 29 October 2013. Speaking on the Today programme for BBC Radio 4, as well as the Business Report for BBC world N...


Mark Bishop on BBC ...

Mark Bishop, Chair of the Study of Artificial Intelligence and the Simulation of Behaviour, appeared on Newsnight to discuss the ethics of ‘killer robots’. He was approached to give his view on a report raising questions on the et...


AISB YouTube Channel

The AISB has launched a YouTube channel: ( The channel currently holds a number of videos from the AISB 2010 Convention. Videos include the AISB round t...


Lighthill Debates

The Lighthill debates from 1973 are now available on YouTube. You need to a flashplayer enabled browser to view this YouTube video  



AISB event Bulletin Item

LECTURE: "Gesture-sound interaction in digital media", 30 March 2011, Goldsmith's College, LONDON

by Frederic Bevilacqua, Head of the Real-Time Musical Interactions Team IRCAM- Centre Pompidou, STMS-CNRS UPMC Paris, France Location: LG01, New Academic Building Department: Computing Time: 30 March 2011, 14:00 - 15:00

I will present an overview of the research and applications performed by the Real-Time Musical 
Interactions Team of IRCAM (Paris). We have developed for the last seven years various methods
and tools for computer-based gesture analysis, with the general goal to use body movements to 
interact with sonic and/or visual environments. This research has largely been influenced by 
sustained collaborations with musicians/composers and dancers/choreographers. We will present 
some of these works, focusing on gesture research and interfaces. In particular, we will present
the cases of musical interfaces and various experiments we have been carried on in music pedagogy.
We will also present dance performances and interactive installations we have collaborated on. 

In music, we studied physical gestures of musicians such as the bow movement of violin players. 
This allowed us to formalize key concepts about continuous gesture control, gesture vocabulary 
and co-articulation (similarly to speech production). This fundamental research led us to design 
augmented instruments, incorporating these challenging concepts. In parallel, we are designing 
new interfaces and paradigms to control sonic environments, individually or collectively. In 
particular, we are developing tools to re-perform sound and music with such interfaces. 
In particular, we developed a "gesture follower" system that allows for the recognition and 
synchronization of gestures with sound materials. 

In dance, we will present performances and installations, where we used the same technology than
for music. While designed with different goals and aesthetics, two of them use a similar interaction 
principle: the visitor is invited to dance imitating dance material displayed on a large screen. 
This brings us back to open questions with musical interfaces: how can we learn gestures and the 
interaction with digital media, and how this affects our gesture and sound perception?