AliGraph: A Comprehensive Graph Neural Network Platform

@article{Zhu2019AliGraphAC,
  title={AliGraph: A Comprehensive Graph Neural Network Platform},
  author={Rong Zhu and Kun Zhao and Hongxia Yang and Wei Lin and Chang Zhou and Baole Ai and Yong Li and Jingren Zhou},
  journal={Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery \& Data Mining},
  year={2019}
}
  • Rong Zhu, Kun Zhao, Jingren Zhou
  • Published 23 February 2019
  • Computer Science
  • Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
An increasing number of machine learning tasks require dealing with large graph datasets, which capture rich and complex relation- ship among potentially billions of elements. Graph Neural Network (GNN) becomes an effective way to address the graph learning problem by converting the graph data into a low dimensional space while keeping both the structural and property information to the maximum extent and constructing a neural network for training and referencing. However, it is challenging to… 
AliGraph: An Industrial Graph Neural Network Platform
TLDR
AliGraph empowers end-to-end solutions for users to address their scenarios by taking graph into deep learning frameworks by co-designing the flexible Gremlinlike interfaces for both graph query and sampling.
DistDGL: Distributed Graph Neural Network Training for Billion-Scale Graphs
  • Da Zheng, Chao Ma, G. Karypis
  • Computer Science
    2020 IEEE/ACM 10th Workshop on Irregular Applications: Architectures and Algorithms (IA3)
  • 2020
TLDR
The results show that DistDGL achieves linear speedup without compromising model accuracy and requires only 13 seconds to complete a training epoch for a graph with 100 million nodes and 3 billion edges on a cluster with 16 machines.
AGL: A Scalable System for Industrial-purpose Graph Machine Learning
TLDR
AGL is designed, a scalable, fault-tolerance and integrated system, with fully-functional training and inference for GNNs, implemented on mature infrastructures such as MapReduce.
Algorithm and System Co-design for Efficient Subgraph-based Graph Representation Learning
TLDR
A novel framework SUREL is proposed for scalable SGRL by co-designing the learning algorithm and its system support and significantly outperforms the state-of-the-art baseline in both prediction accuracy and efficiency.
Graph Neural Network Training with Data Tiering
TLDR
This work provides a method to statistical analyze and identify more frequently accessed data ahead of GNN training and provides a new data placement and access strategy to further minimize the CPU-GPU communication overhead.
Computing Graph Neural Networks: A Survey from Algorithms to Accelerators
TLDR
A review of the field of GNNs is presented from the perspective of computing, and an in-depth analysis of current software and hardware acceleration schemes is provided, from which a hardware-software, graph-aware, and communication-centric vision for GNN accelerators is distilled.
CogDL: A Toolkit for Deep Learning on Graphs
TLDR
CogDL, an extensive toolkit for deep learning on graphs that allows researchers and developers to easily conduct experiments and build applications, is introduced and the effectiveness of CogDL for real-world applications in AMiner, a large academic mining system is demonstrated.
Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine
TLDR
It is shown that out-of-core pipelined mini-batch training in a single machine outperforms resource-hungry multi-GPU solutions and Marius++, a system for training GNNs over billion-scale graphs, is introduced.
Seastar: vertex-centric programming for graph neural networks
TLDR
The Seastar system is proposed, which presents a vertex-centric programming model for GNN training on GPU and provides idiomatic python constructs to enable easy development of novel homogeneous and heterogeneous GNN models.
A Review on Graph Neural Network Methods in Financial Applications
TLDR
This work provides a comprehensive review of GNN models in recent financial context, and summarizes the GNN methodology for each graph type, application in each area, and proposes some potential research areas.
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 72 REFERENCES
Representation Learning for Attributed Multiplex Heterogeneous Network
TLDR
Results of the offline A/B tests on product recommendation further confirm the effectiveness and efficiency of the framework in practice, and the theoretical analysis of the proposed framework gives its connection with previous works and proving its better expressiveness.
A Comprehensive Survey of Graph Embedding: Problems, Techniques, and Applications
TLDR
This survey conducts a comprehensive review of the literature in graph embedding and proposes two taxonomies ofGraph embedding which correspond to what challenges exist in differentgraph embedding problem settings and how the existing work addresses these challenges in their solutions.
Representation Learning on Graphs: Methods and Applications
TLDR
A conceptual review of key advancements in this area of representation learning on graphs, including matrix factorization-based methods, random-walk based algorithms, and graph neural networks are provided.
Heterogeneous Network Embedding via Deep Architectures
TLDR
It is demonstrated that the rich content and linkage information in a heterogeneous network can be captured by a multi-resolution deep embedding function, so that similarities among cross-modal data can be measured directly in a common embedding space.
Accelerated Attributed Network Embedding
TLDR
An accelerated attributed network embedding algorithm AANE is proposed, which enables the joint learning process to be done in a distributed manner by decomposing the complex modeling and optimization into many sub-problems.
One-Shot Relational Learning for Knowledge Graphs
TLDR
This work proposes a one-shot relational learning framework, which utilizes the knowledge distilled by embedding models and learns a matching metric by considering both the learned embeddings and one-hop graph structures.
LINE: Large-scale Information Network Embedding
TLDR
A novel network embedding method called the ``LINE,'' which is suitable for arbitrary types of information networks: undirected, directed, and/or weighted, and optimizes a carefully designed objective function that preserves both the local and global network structures.
Is a Single Vector Enough?: Exploring Node Polysemy for Network Embedding
TLDR
This paper proposes a polysemous embedding approach for modeling multiple facets of nodes, as motivated by the phenomenon of word polysemy in language modeling, which is adaptive to various existing embedding models, without significantly complicating the optimization process.
metapath2vec: Scalable Representation Learning for Heterogeneous Networks
TLDR
Two scalable representation learning models, namely metapath2vec and metapATH2vec++, are developed that are able to not only outperform state-of-the-art embedding models in various heterogeneous network mining tasks, but also discern the structural and semantic correlations between diverse network objects.
...
1
2
3
4
5
...