The paper deals with the design of an intelligent user interface augmenting the user experience in a museum domain, by providing and immersive audio environment. We focus on the issues concerning multimodal interaction, by taking into account auralvisual perception principles. In addition we highlight the potential of augmenting the visual real environment in a personalized way, thanks to context modeling techniques. The LISTEN project, a system for an immersive audio augmented environment applied in the art exhibition domain, provides an example of modeling and personalization methods affecting the audio interface in terms of content and organization. Categories and Subject Descriptors