#### Filter Results:

#### Publication Year

1975

2016

#### Publication Type

#### Co-author

#### Publication Venue

#### Key Phrases

Learn More

The 2-way quantum finite automaton introduced by Kondacs and Watrous[KW97] can accept non-regular languages with bounded error in polynomial time. If we restrict the head of the automaton to moving classically and to moving only in one direction, the acceptance power of this 1-way quantum finite automaton is reduced to a proper subset of the regular… (More)

We consider the problem of deciding whether a polygonal knot in 3-dimensional Euclidean space is unknotted, ie., capable of being continuously deformed without self-intersection so that it lies in a plane. We show that this problem, UNKNOTTING PROBLEM is in NP. We also consider the problem, SPLITTING PROBLEM of determining whether two or more such polygons… (More)

Extendible hashing is a new access technique, in which the user is guaranteed no more than two page faults to locate the data associated with a given unique identifier, or key. Unlike conventional hashing, extendible hashing has a dynamic structure that grows and shrinks gracefully as the database grows and shrinks. This approach simultaneously solves the… (More)

It is shown that a probabilistic Turing acceptor or transducer running within space bound S can be simulated by a time S 2 parallel machine and therefore by a space S 2 deterministic machine. (Previous simulations ran in space $6.) In order to achieve these simulations, known algorithms are extended for the computation of determinants in small arithmetic… (More)

We present upper bounds for sorting and selecting the median in a fixed number of rounds. These bounds match the known lower bounds to within logarithmic factors. They also have the merit of being "explicit modulo expansion"; that is, probabilistic arguments are used only to obtain expanding graphs, and when explicit constructions for such graphs are found,… (More)

It is shown that if formulas are used to compute Boolean functions in the presence of randomly occurring failures (as has been suggested by von Neumann and others), then 1) there is a limit shictly less than 1/2 to the failure probability per gate that can be tolerated, and 2) formulas that tolerate failures must be deeper (and, therefore, compute more… (More)