Corpus ID: 217494802

TRECVID 2016: Evaluating Video Search, Video Event Detection, Localization, and Hyperlinking

@inproceedings{Awad2016TRECVID2E,
  title={TRECVID 2016: Evaluating Video Search, Video Event Detection, Localization, and Hyperlinking},
  author={G. Awad and J. Fiscus and David Joy and M. Michel and A. Smeaton and Wessel Kraaij and G. Qu{\'e}not and Maria Eskevich and R. Aly and R. Ordelman and M. Ritter and G. Jones and B. Huet and M. Larson},
  booktitle={TRECVID},
  year={2016}
}
HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou… Expand
TRECVID 2017: Evaluating Ad-hoc and Instance Video Search, Events Detection, Video Captioning and Hyperlinking
TLDR
Evaluating Ad-hoc and Instance Video Search, Events Detection, Video Captioning, and Hyperlinking George Awad, Asad Butt, Jonathan Fiscus, David Joy, Andrew Delgado, Willie Mcclinton, Martial Michel, Alan Smeaton, Yvette Graham, Wessel Kraaij, et al. Expand
TRECVID 2019: An evaluation campaign to benchmark Video Activity Detection, Video Captioning and Matching, and Video Search & retrieval
TLDR
This paper is an introduction to the evaluation framework, tasks, data, and measures used in the TRECVID 2019, a TREC-style video analysis and retrieval evaluation. Expand
Multimodal Video-to-Video Linking: Turning to the Crowd for Insight and Evaluation
TLDR
The latest edition of the Video Hyperlinking (LNK) task at TRECVid 2016 is described, and insights into how people understand the relevance relationship between two linked video segments are given. Expand
University of Amsterdam and Renmin University at TRECVID 2016: Searching Video, Detecting Events and Describing Video
TLDR
The 2016 edition of the TRECVID benchmark has been a fruitful participation for the joint-team, resulting in the best overall result for zeroand few-example event detection as well as video description by matching and in generative mode. Expand
TRECVID 2020: A comprehensive campaign for evaluating video retrieval tasks across multiple application domains
TLDR
An introduction to the evaluation framework, tasks, data, and measures used in the TRECVID evaluation campaign is introduced. Expand
IRISA at TrecVid 2017: Beyond Crossmodal and Multimodal Models for Video Hyperlinking
TLDR
The runs that were submitted to the TRECVid Challenge 2017 for the Video Hyperlinking task show a gain in performance over the baseline BiDNN model both when the metadata filter was used and when the keyframe fusion was done with a pseudo-inverse. Expand
Indexed Dataset from YouTube for a Content-Based Video Search Engine
Numerous researches on content-based video indexing and retrieval besides video search engines are tied to a large-scaled video dataset. Unfortunately, reduction in open-sourced datasets resulted inExpand
Neighbourhood Structure Preserving Cross-Modal Embedding for Video Hyperlinking
TLDR
The empirical insights are shared on a number of issues in cross-modal learning, including the preservation of neighbourhood structure in embedding, model fine-tuning and issue of missing modality, for video hyperlinking. Expand
Video Retrieval of Specific Persons in Specific Locations
This thesis explores good practices for improving the detection of specific people in specific places, as defined in the TRECVID 2016 benchmark for instance search in large scale video datasets. AnExpand
Multimodal Video Retrieval with the 2017 IMOTION System
TLDR
This demo will present the 2017 IMOTION system deployed on the IACC.3 collection consisting of 600 hours of Internet Archive video, which was also used in the TRECVID 2016 Ad-Hoc Video Search and in the 2017 Video Browser Showdown (VBS) challenge in which IMOTIONS ranked first. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 24 REFERENCES
Multimodal Video-to-Video Linking: Turning to the Crowd for Insight and Evaluation
TLDR
The latest edition of the Video Hyperlinking (LNK) task at TRECVid 2016 is described, and insights into how people understand the relevance relationship between two linked video segments are given. Expand
MSR-VTT: A Large Video Description Dataset for Bridging Video and Language
TLDR
A detailed analysis of MSR-VTT in comparison to a complete set of existing datasets, together with a summarization of different state-of-the-art video-to-text approaches, shows that the hybrid Recurrent Neural Networkbased approach, which combines single-frame and motion representations with soft-attention pooling strategy, yields the best generalization capability on this dataset. Expand
TRECVid Semantic Indexing of Video: A 6-year Retrospective
TLDR
The data, protocol and metrics used for the main and the secondary tasks, the results obtained and the main approaches used by participants are described. Expand
TRECVID 2015 - An Overview of the Goals, Tasks, Data, Evaluation Mechanisms and Metrics
The TREC Video Retrieval Evaluation (TRECVID) 2011 was a TREC-style video analysis and retrieval evaluation, the goal of which remains to promote progress in content-based exploitation of digitalExpand
CIDEr: Consensus-based image description evaluation
TLDR
A novel paradigm for evaluating image descriptions that uses human consensus is proposed and a new automated metric that captures human judgment of consensus better than existing metrics across sentences generated by various sources is evaluated. Expand
Blip10000: a social video dataset containing SPUG content for tagging and retrieval
TLDR
This work presents a dataset that contains comprehensive semi-professional user-generated (SPUG) content, including audiovisual content, user-contributed metadata, automatic speech recognition transcripts, automatic shot boundary files, and social information for multiple 'social levels'. Expand
Creating HAVIC: Heterogeneous Audio Visual Internet Collection
TLDR
The HAVIC (Heterogeneous Audio Visual Internet Collection) Corpus will ultimately consist of several thousands of hours of unconstrained user-generated multimedia content, designed with an eye toward providing increased challenges for both acoustic and video processing technologies. Expand
Feature-based video key frame extraction for low quality video sequences
We present an approach to key frame extraction for structuring user generated videos on video sharing websites (e. g. YouTube). Our approach is intended to link existing image search engines to videoExpand
The TRECVid 2008 Event Detection evaluation
TLDR
The event detection evaluation was organized to address detection of a set of specific events that would be of potential interest to an operator in the surveillance domain. Expand
Adapting Binary Information Retrieval Evaluation Metrics for Segment-based Retrieval Tasks
TLDR
These metrics are described in the context of a task for the hyperlinking of video segments and can in principle be used with any kind of effectiveness measure that uses binary relevance, and for other segment-baed retrieval tasks. Expand
...
1
2
3
...