License: Creative Commons Attribution 3.0 Unported license (CC BY 3.0)
When quoting this document, please refer to the following
DOI: 10.4230/LIPIcs.ITCS.2020.28
URN: urn:nbn:de:0030-drops-117138
URL: http://dagstuhl.sunsite.rwth-aachen.de/volltexte/2020/11713/
Go to the corresponding LIPIcs Volume Portal


Rubinfeld, Ronitt ; Vasilyan, Arsen

Monotone Probability Distributions over the Boolean Cube Can Be Learned with Sublinear Samples

pdf-format:
LIPIcs-ITCS-2020-28.pdf (0.7 MB)


Abstract

A probability distribution over the Boolean cube is monotone if flipping the value of a coordinate from zero to one can only increase the probability of an element. Given samples of an unknown monotone distribution over the Boolean cube, we give (to our knowledge) the first algorithm that learns an approximation of the distribution in statistical distance using a number of samples that is sublinear in the domain.
To do this, we develop a structural lemma describing monotone probability distributions. The structural lemma has further implications to the sample complexity of basic testing tasks for analyzing monotone probability distributions over the Boolean cube: We use it to give nontrivial upper bounds on the tasks of estimating the distance of a monotone distribution to uniform and of estimating the support size of a monotone distribution. In the setting of monotone probability distributions over the Boolean cube, our algorithms are the first to have sample complexity lower than known lower bounds for the same testing tasks on arbitrary (not necessarily monotone) probability distributions.
One further consequence of our learning algorithm is an improved sample complexity for the task of testing whether a distribution on the Boolean cube is monotone.

BibTeX - Entry

@InProceedings{rubinfeld_et_al:LIPIcs:2020:11713,
  author =	{Ronitt Rubinfeld and Arsen Vasilyan},
  title =	{{Monotone Probability Distributions over the Boolean Cube Can Be Learned with Sublinear Samples}},
  booktitle =	{11th Innovations in Theoretical Computer Science Conference (ITCS 2020)},
  pages =	{28:1--28:34},
  series =	{Leibniz International Proceedings in Informatics (LIPIcs)},
  ISBN =	{978-3-95977-134-4},
  ISSN =	{1868-8969},
  year =	{2020},
  volume =	{151},
  editor =	{Thomas Vidick},
  publisher =	{Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik},
  address =	{Dagstuhl, Germany},
  URL =		{https://drops.dagstuhl.de/opus/volltexte/2020/11713},
  URN =		{urn:nbn:de:0030-drops-117138},
  doi =		{10.4230/LIPIcs.ITCS.2020.28},
  annote =	{Keywords: Learning distributions, monotone probability distributions, estimating support size}
}

Keywords: Learning distributions, monotone probability distributions, estimating support size
Collection: 11th Innovations in Theoretical Computer Science Conference (ITCS 2020)
Issue Date: 2020
Date of publication: 06.01.2020


DROPS-Home | Fulltext Search | Imprint | Privacy Published by LZI