Abstract
This paper discusses some issues related to authoring and navigating World Wide Web documents composed of continuous and non-continuous media, based on a video or audio narration to which static or dynamic documents are attached. The discussion stems from a model able to describe synchronization among media elements and media behavior depending on user interaction. A prototype implementation addressing SMIL as a target language is described.