#### Filter Results:

- Full text PDF available (74)

#### Publication Year

1978

2017

- This year (3)
- Last 5 years (39)
- Last 10 years (72)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Data Set Used

#### Key Phrases

Learn More

- Bob Coecke, Mehrnoosh Sadrzadeh, Stephen Clark
- ArXiv
- 2010

We propose a mathematical framework for a unification of the distributional theory of meaning in terms of vector space models, and a compositional theory for grammatical types, for which we rely on the algebra of Pregroups, introduced by Lambek. This mathematical framework enables us to compute the meaning of a well-typed sentence from the meanings of its… (More)

- Edward Grefenstette, Mehrnoosh Sadrzadeh
- EMNLP
- 2011

Modelling compositional meaning for sentences using empirical distributional methods has been a challenge for computational linguists. We implement the abstract categorical model of Coecke et al. (2010) using data from the BNC and evaluate it. The implementation is based on unsupervised learning of matrices for relational words and applying them to the… (More)

We propose a mathematical framework for a unification of the distributional theory of meaning in terms of vector space models, and a compositional theory for grammatical types, namely Lambek’s pregroup semantics. A key observation is that the monoidal category of (finite dimensional) vector spaces, linear maps and the tensor product, as well as any… (More)

We present a model for compositional distributional semantics related to the framework of Coecke et al. (2010), and emulating formal semantics by representing functions as tensors and arguments as vectors. We introduce a new learning method for tensors, generalising the approach of Baroni and Zamparelli (2010). We evaluate it on two benchmark data sets, and… (More)

We provide a comparative study between neural word representations and traditional vector spaces based on cooccurrence counts, in a number of compositional tasks. We use three different semantic spaces and implement seven tensor-based compositional models, which we then test (together with simpler additive and multiplicative approaches) in tasks involving… (More)

- Dimitri Kartsaklis, Mehrnoosh Sadrzadeh, Stephen G. Pulman
- COLING
- 2012

This short paper summarizes a faithful implementation of the categorical framework of Coecke et al. (2010), the aim of which is to provide compositionality in distributional models of lexical semantics. Based on Frobenius Algebras, our method enable us to (1) have a unifying meaning space for phrases and sentences of different structure and word vectors,… (More)

- Edward Grefenstette, Mehrnoosh Sadrzadeh
- ArXiv
- 2011

Formal and distributional semantic models offer complementary benefits in modeling meaning. The categorical compositional distributional model of meaning of Coecke et al. (2010) (abbreviated to DisCoCat in the title) combines aspects of both to provide a general framework in which meanings of words, obtained distributionally, are composed using methods from… (More)

- Minghui Ma, Alessandra Palmigiano, Mehrnoosh Sadrzadeh
- Ann. Pure Appl. Logic
- 2011

In this paper, we start studying epistemic updates using the standard toolkit of duality theory. We focus on public announcements, which are the simplest epistemic actions, and hence on single-agent1 Public Announcement Logic (PAL) without the common knowledge operator. As is well known, the epistemic action of publicly announcing a given proposition is… (More)

- Alexandru Baltag, Bob Coecke, Mehrnoosh Sadrzadeh
- J. Log. Comput.
- 2007

We provide algebraic semantics together with a sound and complete sequent calculus for information update due to epistemic actions. This semantics is flexible enough to accommodate incomplete as well as wrong information e.g. due to secrecy and deceit, as well as nested knowledge. We give a purely algebraic treatment of the muddy children puzzle, which… (More)

- Dimitri Kartsaklis, Mehrnoosh Sadrzadeh
- EMNLP
- 2013

Recent work has shown that compositionaldistributional models using element-wise operations on contextual word vectors benefit from the introduction of a prior disambiguation step. The purpose of this paper is to generalise these ideas to tensor-based models, where relational words such as verbs and adjectives are represented by linear maps (higher order… (More)