Hugo Gonçalo Oliveira

Learn More
In this paper, we present Second HAREM, the second edition of an evaluation campaign for Portuguese, addressing named entity recognition (NER). This second edition also included two new tracks: the recognition and normalization of temporal entities (proposed by a group of participants, and hence not covered on this paper) and ReRelEM, the detection of(More)
This paper describes a project aimed at creating a lexical ontology extracted (semi) automatically from a large Portuguese general dictionary. Although using machine readable dictionaries to extract semantic information is not new, we believe this is the first attempt for the Portuguese language. The paper describes a (to be) freely available resource,(More)
The geographic distribution of genetic diversity and the population structure of tetraploid wheat landraces in the Mediterranean basin has received relatively little attention. This is complicated by the lack of consensus concerning the taxonomy of tetraploid wheats and by unresolved questions regarding the domestication and spread of naked wheats. These(More)
Faba bean (Vicia faba L.) is a facultative cross-pollinating legume crop with a great importance for food and feed due to its high protein content as well as the important role in soil fertility and nitrogen fixation. In this work we evaluated genetic diversity and population structure of faba bean accessions from the Western Mediterranean basin and wild(More)
This paper is about an ongoing project, Tra-la-Lyrics which aims to create a computer program capable of generating lyrics for given melodies. As a preliminary phase, the correlations between words and rhythm in a song’s lyrics was studied and some heuristics were achieved. Algorithms for syllabic division and syllabic stress identification on a word were(More)
In order to deal with ambiguity in natural language, it is common to organise words, according to their senses, in synsets, which are groups of synonymous words that can be seen as concepts. The manual creation of a broad-coverage synset base is a timeconsuming task, so we take advantage of dictionary definitions for extracting synonymy pairs and clustering(More)
Public domain collaborative resources like Wiktionary and Wikipedia have recently become attractive sources for information extraction. To use these resources in natural languague processing (NLP) tasks, efficient programmatic access to their contents is required. In this work, we have extracted semantic relations automatically from the Portuguese(More)
This ongoing research presents an alternative to the manual creation of lexical resources and proposes an approach towards the automatic construction of a lexical ontology for Portuguese. Textual sources are exploited in order to obtain a lexical network based on terms and, after clustering and mapping, a wordnet-like lexical ontology is created. At the end(More)