Visual Systems Laboratory, Intelligent Systems Group, Department of Electronics, University of York |
|
MOTION PICTURE PARSING AND SUMMARIZATION We are interested in parsing video for postproduction and archiving applications. On the one hand we seek frame-accurate descriptions of shot boundaries, frame-approximate descriptions of camera moves, shot-by-shot descriptions of framings and groupings, and a framework for semantic analysis of content. On the other, we want to devise rich static visualizations of shots -- reverse storyboards -- that allow the story of the shot to be told economically and unambiguously. SAM At the lowest level we use a real-time mosaicing method called Simplex Adapted Mesh (SAM) [*] that estimates frame-to-frame projective transforms robustly. This is applied within a causal dual-mosaic framework [*] that allows the full shot mosaic to be built without revisiting inaccurate interframe estimates. SAM was first realized in a system for real-time interactive image mosaicing.
ASAP In research on automated movie analysis[*], we have developed an Automated Shot Analysis Program (ASAP) that parses movies in a way that is useful for people in the production industry. Because its underlying technology is SAM ASAP has no problem handling black-and-white footage, cartoons, or any other kind of source material.
SALSA (separate page) is a semi-automated extension to ASAP suitable for use in postproduction and archiving. Reverse Storyboarding We are continuing to research reverse storyboarding methods [*] [*] for video summarization. This work, done in collaboration with Bob Dony of the University of Guelph, has yielded visualizations that combine storyboard metaphors like onion skins and streak lines with framings and arrows. The top example here shows a relatively simple shot with object and camera motion in the same direction. The bottom example is much more complicated with several people moving in different ways within a panning shot. |