Von Neumann type trace inequalities for Schatten-class operators

  title={Von Neumann type trace inequalities for Schatten-class operators},
  author={Gunther Dirr and Frederik vom Ende},
  journal={Journal of Operator Theory},
We generalize von Neumann's well-known trace inequality, as well as related eigenvalue inequalities for hermitian matrices, to Schatten-class operators between complex Hilbert spaces of infinite dimension. To this end, we exploit some recent results on the $C$-numerical range of Schatten-class operators. For the readers' convenience, we sketched the proof of these results in the Appendix. 

Reachability in Controlled Markovian Quantum Systems: An Operator-Theoretic Approach

In quantum systems theory one of the fundamental problems boils down to: Given an initial state, which final states can be reached by the dynamic system in question? Formulated in the framework of

A trace inequality of John von Neumann

AbstractThe principal aim of this note is to establish an effectively self-contained proof of J. von Neumann's inequality $$\left| {tr\left( {AB} \right)} \right| \leqslant \sum\limits_{r = 1}^n

The C-Numerical Range in Infinite Dimensions

ABSTRACT In infinite dimensions and on the level of trace-class operators C rather than matrices, we show that the closure of the C-numerical range is always star-shaped with respect to the set ,

The C-numerical range of matrices is star-shaped

Let A C be n× ncomplex matrices. We prove in the affirmative the conjecture that the C-numerical range of A, defined by is always star-shaped with respect to star-center (tr A)(tr C)/ n. This result

Another proof of a result of Westwick

Let be the set of all n × n unitary matrices and A, C two n × n complex matrices with C hermitian. In this note, the author gives another proof of the following result: the set is convex.

Unitary control in quantum ensembles: maximizing signal intensity in coherent spectroscopy

A gradient-based systematic procedure for optimizing these transformations is described that finds the largest projection of a transformed initial operator onto the target operator and, thus, the maximum spectroscopic signal.

Neural Networks, Principal Components, and Subspaces

    E. Oja
    Computer Science
    Int. J. Neural Syst.
  • 1989
A single neuron with Hebbian-type learning for the connection weights, and with nonlinear internal feedback, has been shown to extract the statistical principal components of its stationary input pattern sequence, which yields a multi-dimensional, principal component subspace.

A theory of cross-spaces

Neural networks