#### Filter Results:

- Full text PDF available (163)

#### Publication Year

1995

2018

- This year (5)
- Last 5 years (83)
- Last 10 years (154)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Data Set Used

#### Key Phrases

Learn More

- Ümit V. Çatalyürek, Cevdet Aykanat
- IEEE Trans. Parallel Distrib. Syst.
- 1999

ÐIn this work, we show that the standard graph-partitioning-based decomposition of sparse matrices does not reflect the actual communication volume requirement for parallel matrix-vector multiplication. We propose two computational hypergraph models which avoid this crucial deficiency of the graph model. The proposed models reduce the decomposition problem… (More)

Many emerging large-scale data science applications require searching large graphs distributed across multiple memories and processors. This paper presents a distributed breadthfirst search (BFS) scheme that scales for random graphs with up to three billion vertices and 30 billion edges. Scalability was tested on IBM BlueGene/L with 32,768 nodes at the… (More)

- Karen D. Devine, Erik G. Boman, Robert T. Heaphy, Rob H. Bisseling, Ümit V. Çatalyürek
- Proceedings 20th IEEE International Parallel…
- 2006

Graph partitioning is often used for load balancing in parallel computing, but it is known that hypergraph partitioning has several advantages. First, hypergraphs more accurately model communication volume, and second, they are more expressive and can better represent nonsymmetric problems. Hypergraph partitioning is particularly suited to parallel sparse… (More)

- Ümit V. Çatalyürek, Cevdet Aykanat
- Proceedings 15th International Parallel and…
- 2001

We propose a new hypergraph model for the decomposition of irregular computational domains. This work focuses on the decomposition of sparse matrices for parallel matrix-vector multiplication. However, the proposed model can also be used to decompose computational domains of other parallel reduction problems. We propose a “finegrain” hypergraph model for… (More)

- Cevdet Aykanat, Ali Pinar, Ümit V. Çatalyürek
- SIAM J. Scientific Computing
- 2004

We investigate the problem of permuting a sparse rectangular matrix into blockdiagonal form. Block-diagonal form of a matrix grants an inherent parallelism for solving the deriving problem, as recently investigated in the context of mathematical programming, LU factorization, and QR factorization. To represent the nonzero structure of a matrix, we propose… (More)

- Ümit V. Çatalyürek, Cevdet Aykanat
- Encyclopedia of Parallel Computing
- 2011

- Doruk Bozdag, Ümit V. Çatalyürek, Füsun Özgüner
- Proceedings 20th IEEE International Parallel…
- 2006

We propose a new duplication-based DAG scheduling algorithm for heterogeneous computing environments. Contrary to the traditional approaches, proposed algorithm traverses the DAG in a bottom-up fashion while taking advantage of task duplication and task insertion. Experimental results on random DAGs and three different application DAGs show that the… (More)

- A RECIPE, Ümit V. Çatalyürek, Bora Uçar
- 2011

We consider two-dimensional partitioning of general sparse matrices for parallel sparse matrix-vector multiply operation. We present three hypergraph-partitioning-based methods, each having unique advantages. The first one treats the nonzeros of the matrix individually and hence produces fine-grain partitions. The other two produce coarser partitions, where… (More)

- Ümit V. Çatalyürek, Cevdet Aykanat
- IRREGULAR
- 1996

In this work we show the de ciencies of the graph model for decomposing sparse matrices for parallel matrix vector multiplica tion Then we propose two hypergraph models which avoid all de cien cies of the graph model The proposed models reduce the decomposition problem to the well known hypergraph partitioning problem widely en countered in circuit… (More)

- Erik Saule, Kamer Kaya, Ümit V. Çatalyürek
- PPAM
- 2013

Intel Xeon Phi is a recently released high-performance coprocessor which features 61 cores each supporting 4 hardware threads with 512-bit wide SIMD registers achieving a peak theoretical performance of 1Tflop/s in double precision. Many scientific applications involve operations on large sparse matrices such as linear solvers, eigensolver, and graph mining… (More)