Deep Code Comment Generation
- Xing Hu, Ge Li, Xin Xia, D. Lo, Zhi Jin
- Computer ScienceIEEE International Conference on Program…
- 1 May 2018
DeepCom applies Natural Language Processing (NLP) techniques to learn from a large code corpus and generates comments from learned features for better comments generation of Java methods.
Convolutional Neural Networks over Tree Structures for Programming Language Processing
- Lili Mou, Ge Li, Lu Zhang, Tao Wang, Zhi Jin
- Computer ScienceAAAI Conference on Artificial Intelligence
- 18 September 2014
A novel tree-based convolutional neural network (TBCNN) is proposed for programming language processing, in which a convolution kernel is designed over programs' abstract syntax trees to capture structural information.
Classifying Relations via Long Short Term Memory Networks along Shortest Dependency Paths
- Yan Xu, Lili Mou, Ge Li, Yunchuan Chen, Hao Peng, Zhi Jin
- Computer ScienceConference on Empirical Methods in Natural…
- 15 August 2015
This paper presents SDP-LSTM, a novel neural network to classify the relation of two entities in a sentence, which leverages the shortest dependency path (SDP) between two entities; multichannel recurrent neural networks, with long short term memory (L STM) units, pick up heterogeneous information along the SDP.
Summarizing Source Code with Transferred API Knowledge
- Xing Hu, Ge Li, Xin Xia, D. Lo, Shuai Lu, Zhi Jin
- Computer ScienceInternational Joint Conference on Artificial…
- 1 July 2018
Experiments on large-scale real-world industry Java projects indicate that the proposed novel approach, named TL-CodeSum, is effective and outperforms the state-of-the-art in code summarization.
Natural Language Inference by Tree-Based Convolution and Heuristic Matching
This model, a tree-based convolutional neural network (TBCNN) captures sentence-level semantics; then heuristic matching layers like concatenation, element-wise product/difference combine the information in individual sentences.
How Transferable are Neural Networks in NLP Applications?
- Lili Mou, Zhao Meng, Zhi Jin
- Computer ScienceConference on Empirical Methods in Natural…
- 19 March 2016
In this paper, systematic case studies are conducted and an illuminating picture is provided on the transferability of neural networks in NLP.
Sequence to Backward and Forward Sequences: A Content-Introducing Approach to Generative Short-Text Conversation
- Lili Mou, Yiping Song, Rui Yan, Ge Li, Lu Zhang, Zhi Jin
- Computer ScienceInternational Conference on Computational…
- 1 July 2016
This paper proposes seq2BF, a “sequence to backward and forward sequences” model, which generates a reply containing the given keyword, and significantly outperforms traditional sequence-to-sequence models in terms of human evaluation and the entropy measure.
Deep code comment generation with hybrid lexical and syntactical information
Experimental results demonstrate that the method Hybrid-DeepCom outperforms the state-of-the-art by a substantial margin and the results show that reducing the out- of-vocabulary tokens improves the accuracy effectively.
Detecting Code Clones with Graph Neural Network and Flow-Augmented Abstract Syntax Tree
- Wenhan Wang, Ge Li, Bo Ma, Xin Xia, Zhi Jin
- Computer ScienceIEEE International Conference on Software…
- 1 February 2020
The first to apply graph neural networks on the domain of code clone detection and build a graph representation of programs called flow-augmented abstract syntax tree (FA-AST), which outperforms the state-of-the-art approaches on both Google Code Jam and BigCloneBench tasks.
Code Generation as a Dual Task of Code Summarization
- Bolin Wei, Ge Li, Xin Xia, Zhiyi Fu, Zhi Jin
- Computer ScienceNeural Information Processing Systems
- 1 October 2019
This paper proposes a dual training framework to train the two tasks simultaneously, and considers the dualities on probability and attention weights, and design corresponding regularization terms to constrain the duality.
...
...