Learn More
We are currently investigating methods to triplify the content of Wikipedia's tables. We propose that existing knowledge-bases can be leveraged to semi-automatically extract high-quality facts (in the form of RDF triples) from tables embedded in Wikipedia articles (henceforth called " Wikitables "). We present a survey of Wikitables and their content in a(More)
The tables embedded in Wikipedia articles contain rich, semi-structured encyclopaedic content. However, the cumulative content of these tables cannot be queried against. We thus propose methods to recover the semantics of Wikipedia tables and, in particular, to extract facts from them in the form of RDF triples. Our core method uses an existing Linked Data(More)
Antigens (Ags) are converted into immune complexes (antigen-antibody complexes, IC) as soon as they encounter their specific antibodies (Abs). In fluids containing complement, the process of IC formation and fixation of complement components occur simultaneously. Hence, the formation of Ag-Ab-complement complexes is the normal way of eliminating Ags from a(More)
An efficient implementation of an algorithm that decides the implication problem for a tractable and expressive class of XML keys. Performance Analysis: Implication problem over large sets of XML Keys. Non-redundant covers of XML keys and validation of XML documents. Our experiments show that reasoning about expressive notions of XML keys can be done(More)
In this paper, we describe our contribution to the 2015 Linked Data Mining Challenge. The proposed task is concerned with the prediction of review of movies as " good " or " bad " , as does Meta-critic website based on critics' reviews. First we describe the sources used to build the training data. Although, several sources provide data about movies on the(More)
Linked Data (LD) datasets (e.g., DBpedia, Freebase) are used in many knowledge extraction tasks due to the high variety of domains they cover. Unfortunately, many of these datasets do not provide a description for their properties and classes, reducing the users' freedom to understand, reuse or enrich them. This work attempts to fill part of this lack by(More)
This paper describes µRaptor, a DOM-based method to extract hCard microformats from HTML pages stripped of microformat markup. µRaptor extracts DOM sub-trees, converts them into rules, and uses them to extract hCard microformats. Besides, we use co-occurring CSS classes to improve the overall precision. Results on train data show 0.96 precision and 0.83 F1(More)
Tables are widely used in Wikipedia articles to display re-lational information – they are inherently concise and information rich. However, aside from info-boxe s, there are no automatic methods to exploit the integrated content of these tables. We thus present DRETa: a tool that uses DBpedia as a reference knowledge-base to extract RDF triples from(More)
We introduce soft cardinality constraints which need to be satisfied on average only, and thus permit violations in a controlled manner. Starting from a highly expressive but intractable class, we establish a fragment that is maximal with respect to both expressivity and efficiency. More precisely, we characterise the associated implication problem(More)