๐ฎ
๐ฎ
The Ethereal
Average Bias and Polynomial Sources
May 28, 2019 ยท The Ethereal ยท ๐ Electron. Colloquium Comput. Complex.
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Arnab Bhattacharyya, Philips George John, Suprovat Ghoshal, Raghu Meka
arXiv ID
1905.11612
Category
cs.CC: Computational Complexity
Cross-listed
cs.DM,
cs.DS
Citations
1
Venue
Electron. Colloquium Comput. Complex.
Last Checked
2 months ago
Abstract
We identify a new notion of pseudorandomness for randomness sources, which we call the average bias. Given a distribution $Z$ over $\{0,1\}^n$, its average bias is: $b_{\text{av}}(Z) =2^{-n} \sum_{c \in \{0,1\}^n} |\mathbb{E}_{z \sim Z}(-1)^{\langle c, z\rangle}|$. A source with average bias at most $2^{-k}$ has min-entropy at least $k$, and so low average bias is a stronger condition than high min-entropy. We observe that the inner product function is an extractor for any source with average bias less than $2^{-n/2}$. The notion of average bias especially makes sense for polynomial sources, i.e., distributions sampled by low-degree $n$-variate polynomials over $\mathbb{F}_2$. For the well-studied case of affine sources, it is easy to see that min-entropy $k$ is exactly equivalent to average bias of $2^{-k}$. We show that for quadratic sources, min-entropy $k$ implies that the average bias is at most $2^{-ฮฉ(\sqrt{k})}$. We use this relation to design dispersers for separable quadratic sources with a min-entropy guarantee.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computational Complexity
๐ฎ
๐ฎ
The Ethereal
An Exponential Separation Between Randomized and Deterministic Complexity in the LOCAL Model
๐ฎ
๐ฎ
The Ethereal
The Parallelism Tradeoff: Limitations of Log-Precision Transformers
๐ฎ
๐ฎ
The Ethereal
The Hardness of Approximation of Euclidean k-means
๐ฎ
๐ฎ
The Ethereal
Slightly Superexponential Parameterized Problems
๐ฎ
๐ฎ
The Ethereal