We study the problem of certification: given queries to a function f : {0,1}n β {0,1} with certificate complexity β€ k and an input xβ, output a size-k certificate for fβs value on xβ. For monotone functions, a classic local search algorithm of Angluin accomplishes this task with n queries, which we show is optimal for local search algorithms. Our main result is a new algorithm for certifying monotone functions with O(k8 logn) queries, which comes close to matching the information-theoretic lower bound of Ξ©(k logn). The design and analysis of our algorithm are based on a new connection to threshold phenomena in monotone functions. We further prove exponential-in-k lower bounds when f is non-monotone, and when f is monotone but the algorithm is only given random examples of f. These lower bounds show that assumptions on the structure of f and query access to it are both necessary for the polynomial dependence on k that we achieve.
more »
« less
Certification with an NP oracle
In the certification problem, the algorithm is given a function f with certificate complexity k and an input x^β, and the goal is to find a certificate of size β€ poly(k) for fβs value at x^β. This problem is in NP^NP, and assuming π― β NP, is not in π―. Prior works, dating back to Valiant in 1984, have therefore sought to design efficient algorithms by imposing assumptions on f such as monotonicity. Our first result is a BPP^NP algorithm for the general problem. The key ingredient is a new notion of the balanced influence of variables, a natural variant of influence that corrects for the bias of the function. Balanced influences can be accurately estimated via uniform generation, and classic BPP^NP algorithms are known for the latter task. We then consider certification with stricter instance-wise guarantees: for each x^β, find a certificate whose size scales with that of the smallest certificate for x^β. In sharp contrast with our first result, we show that this problem is NP^NP-hard even to approximate. We obtain an optimal inapproximability ratio, adding to a small handful of problems in the higher levels of the polynomial hierarchy for which optimal inapproximability is known. Our proof involves the novel use of bit-fixing dispersers for gap amplification.
more »
« less
- Award ID(s):
- 2006664
- PAR ID:
- 10434685
- Date Published:
- Journal Name:
- Proceedings of the 14th Innovations in Theoretical Computer Science Conference (ITCS 2023)
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Existing proofs that deduce BPP = P from circuit lower bounds convert randomized algorithms into deterministic algorithms with a large polynomial slowdown. We convert randomized algorithms into deterministic ones with little slowdown . Specifically, assuming exponential lower bounds against randomized NP β© coNP circuits, formally known as randomized SVN circuits, we convert any randomized algorithm over inputs of length n running in time t β₯ n into a deterministic one running in time t 2+Ξ± for an arbitrarily small constant Ξ± > 0. Such a slowdown is nearly optimal for t close to n , since under standard complexity-theoretic assumptions, there are problems with an inherent quadratic derandomization slowdown. We also convert any randomized algorithm that errs rarely into a deterministic algorithm having a similar running time (with pre-processing). The latter derandomization result holds under weaker assumptions, of exponential lower bounds against deterministic SVN circuits. Our results follow from a new, nearly optimal, explicit pseudorandom generator fooling circuits of size s with seed length (1+Ξ±)log s , under the assumption that there exists a function f β E that requires randomized SVN circuits of size at least 2 (1-Ξ±β²) n , where Ξ± = O (Ξ±)β². The construction uses, among other ideas, a new connection between pseudoentropy generators and locally list recoverable codes.more » « less
-
Evans, Robin; Shpitser, Ilya (Ed.)We consider the problem of maximizing submodular functions under submodular constraints by formulating the problem in two ways: \SCSKC and \DiffC. Given two submodular functions $$f$$ and $$g$$ where $$f$$ is monotone, the objective of \SCSKC problem is to find a set $$S$$ of size at most $$k$$ that maximizes $f(S)$ under the constraint that $$g(S)\leq \theta$$, for a given value of $$\theta$$. The problem of \DiffC focuses on finding a set $$S$$ of size at most $$k$$ such that $h(S) = f(S)-g(S)$$ is maximized. It is known that these problems are highly inapproximable and do not admit any constant factor multiplicative approximation algorithms unless NP is easy. Known approximation algorithms involve data-dependent approximation factors that are not efficiently computable. We initiate a study of the design of approximation algorithms where the approximation factors are efficiently computable. For the problem of \SCSKC, we prove that the greedy algorithm produces a solution whose value is at least $$(1-1/e)f(\OPT) - A$, where $$A$$ is the data-dependent additive error. For the \DiffC problem, we design an algorithm that uses the \SCSKC greedy algorithm as a subroutine. This algorithm produces a solution whose value is at least $$(1-1/e)h(\OPT)-B$, where $$B$$ is also a data-dependent additive error. A salient feature of our approach is that the additive error terms can be computed efficiently, thus enabling us to ascertain the quality of the solutions produced.more » « less
-
null (Ed.)The complexity class ZPP NP[1] (corresponding to zero-error randomized algorithms with access to one NP oracle query) is known to have a number of curious properties. We further explore this class in the settings of time complexity, query complexity, and communication complexity. β’ For starters, we provide a new characterization: ZPP NP[1] equals the restriction of BPP NP[1] where the algorithm is only allowed to err when it forgoes the opportunity to make an NP oracle query. β’ Using the above characterization, we prove a query-to-communication lifting theorem , which translates any ZPP NP[1] decision tree lower bound for a function f into a ZPP NP[1] communication lower bound for a two-party version of f . β’ As an application, we use the above lifting theorem to prove that the ZPP NP[1] communication lower bound technique introduced by GΓΆΓΆs, Pitassi, and Watson (ICALP 2016) is not tight. We also provide a βprimalβ characterization of this lower bound technique as a complexity class.more » « less
-
Aichholzer, Oswin; Wang, Haitao (Ed.)The πβΒ² min-sum k-clustering problem is to partition an input set into clusters C_1,β¦,C_k to minimize β_{i=1}^k β_{p,q β C_i} βp-qββΒ². Although πβΒ² min-sum k-clustering is NP-hard, it is not known whether it is NP-hard to approximate πβΒ² min-sum k-clustering beyond a certain factor. In this paper, we give the first hardness-of-approximation result for the πβΒ² min-sum k-clustering problem. We show that it is NP-hard to approximate the objective to a factor better than 1.056 and moreover, assuming a balanced variant of the Johnson Coverage Hypothesis, it is NP-hard to approximate the objective to a factor better than 1.327. We then complement our hardness result by giving a fast PTAS for πβΒ² min-sum k-clustering. Specifically, our algorithm runs in time O(n^{1+o(1)}dβ 2^{(k/Ξ΅)^O(1)}), which is the first nearly linear time algorithm for this problem. We also consider a learning-augmented setting, where the algorithm has access to an oracle that outputs a label i β [k] for input point, thereby implicitly partitioning the input dataset into k clusters that induce an approximately optimal solution, up to some amount of adversarial error Ξ± β [0,1/2). We give a polynomial-time algorithm that outputs a (1+Ξ³Ξ±)/(1-Ξ±)Β²-approximation to πβΒ² min-sum k-clustering, for a fixed constant Ξ³ > 0.more » « less
An official website of the United States government

