Incremental Learning of Context Free Grammars by Bridging Rule Generation and Search for Semi-optimum Rule Sets

  title={Incremental Learning of Context Free Grammars by Bridging Rule Generation and Search for Semi-optimum Rule Sets},
  author={Katsuhiko Nakamura},
This paper describes novel methods of learning general context free grammars from sample strings, which are implemented in Synapse system. Main features of the system are incremental learning, rule generation based on bottom-up parsing of positive samples, and search for rule sets. From the results of parsing, a rule generation process, called “bridging,” synthesizes the production rules that make up any lacking parts of an incomplete derivation tree for each positive string. To solve the… 
Incremental Learning of Cellular Automata for Parallel Recognition of Formal Languages
Modelling methods to learn the rules of OCAs in the Occam system, which is based on grammatical inference of context-free grammars implemented in Synapse, to develop automatic synthesis of parallel systems and to contribute to the theory of real-time recognition by cellular automata.
Eliminating Stack Symbols in Push-Down Automata and Linear Indexed Grammars
It is shown that the class of languages of PI-LIGs is incomparable with that of PDAs, which is theclass of context-free languages (CFLs), and a simple bottom-up parsing method for LIGs, in which the stack symbols are eliminated at the first step of the parsing.
Towards Machine Learning of Grammars and Compilers of Programming Languages
An application of incremental learning of definite clause grammars (DCGs) and syntax directed translation schema (SDTS) and Synapse synthesized a set of SDTS rules for translating extended arithmetic expressions with function calls and assignment operators into object codes from positive and negative samples of the translation.
Gramin: a system for incremental learning of programming language grammars
A programming language grammar inference system, called Gramin, which is used to infer grammar from sample programs, and employs various optimizations to make grammar inference practical in the domain of programming languages.
Automatic grammar repair
We describe the first approach to automatically repair bugs in context-free grammars: given a grammar that fails some tests in a given test suite, we iteratively and gradually transform the grammar
Large Scale Inference of Deterministic Transductions: Tenjinno Problem 1
A new strategy for inferring large scale transducers that is more adapted for large random instances of the type in question is developed, which involved combining traditional state merging algorithms for inference of finite state automata with EM based alignment algorithms and state splitting algorithms.
Unsupervised Induction of Labeled Parse Trees by Clustering with Syntactic Features
An algorithm for unsupervised induction of labeled parse trees is presented, which obtains 59% labeled f-score on the WSJ10 corpus, as compared to 35% in previous work, and substantial error reduction over a random baseline.
Human behavior recognition using a context-free grammar
This work introduces several important novelties, such as the capability to learn actions based on both positive and negative samples, the possibility of efficiently retraining the system in the presence of misclassified or unrecognized events, and the use of a parsing procedure that allows correct detection of the activities also when they are concatenated and/or nested one with each other.
Inference of k-Testable Directed Acyclic Graph Languages
This paper extends the well-known classes of k-testability and k- testability in the strict sense languages to directed graph languages and proposes a graph automata model for directed acyclic graph languages that is used to propose a grammatical inference algorithm to learn the class of directed acYclic k- Testable in the stricter sense graph languages.


Incremental Learning of Context Free Grammars by Parsing-Based Rule Generation and Rule Set Search
The improved version of Synapse uses a novel rule generation method, called ``bridging,'' which bridges the lacked part of the derivation tree for the positive string, and employs a novel search strategy, called serial search in addition to minimum rule set search.
Incremental learning of context free grammars based on bottom-up parsing and search
Incremental Learning of Context Free Grammars by Extended Inductive CYK Algorithm
The improved version of Synapse employs incremental learning based on the rule generation mechanism called inductive CYK algorithm, which generates the minimum production rules required for parsing positive samples, and the form of production rules is extended to include not only A →βγ but also A → β, called extended Chomsky normal form.
Incremental Learning of Context Free Grammars
In inductive inference for synthesizing context free grammars from positive and negative sample strings, implemented in Synapse system, is described.
Learning Context-Free Grammars from Partially Structured Examples
This paper employs the GA-based learning algorithm for context-free grammars using tabular representations, and presents an algorithm to eliminate unnecessary nonterminals and production rules using the partially structured examples at the initial stage of theGA- based learning algorithm.
Synthesizing Context Free Grammars from Sample Strings Based on Inductive CYK Algorithm
This paper describes a method of synthesizing context free grammars from positive and negative sample strings, which is implemented in a grammatical inference system called Synapse. The method is
Learning Context-Free Grammars with a Simplicity Bias
A rational reconstruction of Wolff's SNPR - the GRIDS system - is presented which incorporates a bias toward grammars that minimize description length, and the algorithm alternates between merging existing nonterminal symbols and creating new symbols, using a beam search to move from complex to simpler Grammars.