Learn More
This paper considers the problem of automatically learning an activity-based semantic scene model from a stream of video data. A scene model is proposed that labels regions according to an identifiable activity in each region, such as entry/exit zones, junctions, paths, and stop zones. We present several unsupervised methods that learn these scene elements(More)
We introduce a large body of virtual human action silhouette (ViHASi) data generated recently for the purpose of evaluating a family of action recognition methods. These are the silhouette-based human action recognition methods. This synthetic multi-camera video data-set consists of 20 action classes, 9 actors and up to 40 synchronized perspective cameras.(More)