"Information Theory and the Complexity of Boolean Functions" by Nicholas Pippenger
 

Information Theory and the Complexity of Boolean Functions

Document Type

Article

Department

Mathematics (HMC)

Publication Date

1976

Abstract

This paper explores the connections between two areas pioneered by Shannon: the transmission of information with a fidelity criterion, and the realization of Boolean functions by networks and formulae. We study three phenomena:

  1. The effect of the relative number of O's and l's in a function's table on its complexity.

  2. The effect of the number of unspecified entries in a partially specified function's table on its complexity.

  3. The effect of the number of errors allowed in the realization of a function on its complexity.

Our main result is a precise version of the following statement:

The complexity of approximately realizing a partially specified Boolean function, in whose table a fractiond of the entries are unspecified and a fractionp of the specified entries are l'swith errors allowed in a fraction not more thane of the specified entries, is less by the factor (1 −d) [H(p) − H(e)] (whereH(z) = −z log2 z −(1 −z) log2 (1 −z) is the binary entropy function) than the complexity of exactly realizing an arbitrary fully specified Boolean function.

We also give an intuitively appealing information-theoretic interpretation of the result.

Rights Information

© 1976 Springer-Verlag

Terms of Use & License Information

Terms of Use for work posted in Scholarship@Claremont.

Share

COinS