Skip to main content
Karl Wimmer
    A function $f$ is $d$-resilient if all its Fourier coefficients of degree at most $d$ are zero, i.e., $f$ is uncorrelated with all low-degree parities. We study the notion of $\mathit{approximate}$ $\mathit{resilience}$ of Boolean... more
    A function $f$ is $d$-resilient if all its Fourier coefficients of degree at most $d$ are zero, i.e., $f$ is uncorrelated with all low-degree parities. We study the notion of $\mathit{approximate}$ $\mathit{resilience}$ of Boolean functions, where we say that $f$ is $\alpha$-approximately $d$-resilient if $f$ is $\alpha$-close to a $[-1,1]$-valued $d$-resilient function in $\ell_1$ distance. We show that approximate resilience essentially characterizes the complexity of agnostic learning of a concept class $C$ over the uniform distribution. Roughly speaking, if all functions in a class $C$ are far from being $d$-resilient then $C$ can be learned agnostically in time $n^{O(d)}$ and conversely, if $C$ contains a function close to being $d$-resilient then agnostic learning of $C$ in the statistical query (SQ) framework of Kearns has complexity of at least $n^{\Omega(d)}$. This characterization is based on the duality between $\ell_1$ approximation by degree-$d$ polynomials and approxim...
    ABSTRACT A function $f:\mathbb{F}_2^n \to \{-1,1\}$ is called linear-isomorphic to g if f=g∘A for some non-singular matrix A. In the g-isomorphism problem, we want a randomized algorithm that distinguishes whether an input function f is... more
    ABSTRACT A function $f:\mathbb{F}_2^n \to \{-1,1\}$ is called linear-isomorphic to g if f=g∘A for some non-singular matrix A. In the g-isomorphism problem, we want a randomized algorithm that distinguishes whether an input function f is linear-isomorphic to g or far from being so. We show that the query complexity to test g-isomorphism is essentially determined by the spectral norm of g. That is, if g is close to having spectral norm s, then we can test g-isomorphism with poly(s) queries, and if g is far from having spectral norm s, then we cannot test g-isomorphism with o(logs) queries. The upper bound is almost tight since there is indeed a function g close to having spectral norm s whereas testing g-isomorphism requires Ω(s) queries. As far as we know, our result is the first characterization of this type for functions. Our upper bound is essentially the Kushilevitz-Mansour learning algorithm, modified for use in the implicit setting. Exploiting our upper bound, we show that any property is testable if it can be well-approximated by functions with small spectral norm. We also extend our algorithm to the setting where A is allowed to be singular.
    ABSTRACT Given an implicit $n\times n$ matrix $A$ with oracle access $x^TA x$ for any $x\in \mathbb{R}^n$, we study the query complexity of randomized algorithms for estimating the trace of the matrix. This problem has many applications... more
    ABSTRACT Given an implicit $n\times n$ matrix $A$ with oracle access $x^TA x$ for any $x\in \mathbb{R}^n$, we study the query complexity of randomized algorithms for estimating the trace of the matrix. This problem has many applications in quantum physics, machine learning, and pattern matching. Two metrics are commonly used for evaluating the estimators: i) variance; ii) a high probability multiplicative-approximation guarantee. Almost all the known estimators are of the form $\frac{1}{k}\sum_{i=1}^k x_i^T A x_i$ for $x_i\in \mathbb{R}^n$ being i.i.d. for some special distribution. Our main results are summarized as follows. We give an exact characterization of the minimum variance unbiased estimator in the broad class of linear nonadaptive estimators (which subsumes all the existing known estimators). We also consider the query complexity lower bounds for any (possibly nonlinear and adaptive) estimators: (1) We show that any estimator requires $\Omega(1/\epsilon)$ queries to have a guarantee of variance at most $\epsilon$. (2) We show that any estimator requires $\Omega(\frac{1}{\epsilon^2}\log \frac{1}{\delta})$ queries to achieve a $(1\pm\epsilon)$-multiplicative approximation guarantee with probability at least $1 - \delta$. Both above lower bounds are asymptotically tight. As a corollary, we also resolve a conjecture in the seminal work of Avron and Toledo (Journal of the ACM 2011) regarding the sample complexity of the Gaussian Estimator.
    In a very strong positive result for passive learn- ing algorithms, Bshouty et al. showed that DNF expressions are efficiently learnable in the uni- form random walk model. It is natural to ask whether the more expressive class of thresh-... more
    In a very strong positive result for passive learn- ing algorithms, Bshouty et al. showed that DNF expressions are efficiently learnable in the uni- form random walk model. It is natural to ask whether the more expressive class of thresh- olds of parities (TOP) is similarly learnable, but the Bshouty et al. time bound becomes exponential in this case. We