Background

One of the most important measures of the complexity of a Boolean function f: ā„ n ā†’{Ā±1} is that of its average sensitivity, namely

š”ø š•Š ( f ) : = š”¼ x āˆ¼ u { Ā± 1 } n # { i : f ( x ) ā‰  f ( x i ) }

where xi above is x with the ith coordinate flipped. The average sensitivity and related measures of noise sensitivity of a Boolean function have found several applications, perhaps most notably to the area of machine learning (see for example [1]). It has thus become important to understand how large the average sensitivity of functions in various classes can be.

Of particular interest is the study of the sensitivity of certain classes of algebraically defined functions. Gotsman and Linial [2] first studied the sensitivity of polynomial threshold functions (i.e. functions of the form f(x)=sgn(p(x)) for p a polynomial of bounded degree). They conjectured exact upper bounds on the sensitivity of polynomial threshold functions of limited degree, but were unable to prove them except in the case of linear threshold functions (when p is required to be degree 1). Since then, significant progress has been made towards proving this Conjecture. The first non-trivial bounds for large degree were proven in [3] by Diakonikolas et al. in 2010. Since then, progress has been rapid. In [4], the Gaussian analogue of the Gotsman-Linial Conjecture was proved, and in [5] the correct bound on average sensitivity was proved to within a polylogarithmic factor.

Another potential generalization of the degree-1 case of the Gotsman-Linial Conjecture (which bounds the sensitivity of the indicator function of a halfspace) would be to consider the sensitivity of the indictor function of the intersection of a bounded number of halfspaces. The Gaussian analogue of this question has already been studied. In particular, Nazarov has shown (see [6]) that the Gaussian surface area of an intersection of k halfspaces is at most O log k . This suggests that the average sensitivity of such a function should be bounded by O n log k . Although this bound has been believed for some time, attempts to prove it have been unsuccessful. Perhaps the closest attempt thus far was by Harsha, Klivans and Meka who show in [7] that an intersection of k sufficiently regular halfspaces has noise sensitivity with parameter Īµ at most log(k)O(1)Īµ1/6. In this paper, we prove that the bound of O n log ( k ) is in face correct. In particular, we prove the following Theorem:

Theorem 1.

Let f be the indicator function of an intersection of k half spaces in n variables, then

š”ø š•Š ( f ) = O n log ( k ) .

It should also be noted that Nazarovā€™s bound follows as a Corollary of Theorem 1, by replacing Gaussian random variables with averages of Bernoulli random variables. It is also not hard to show that this bound is tight up to constants. In particular:

Theorem 2.

If kā‰¤2n, there exists a function f in n variables given by the intersection of at most k half spaces so that

š”ø š•Š ( f ) = Ī© n log ( k ) .

Our proof of Theorem 1 actually uses very little information about halfspaces. In particular, we use only the fact that linear threshold functions are monotonic in the following sense:

Definition 1.

We say that a function f: { Ā± 1 } n ā†’ā„ is unate if for all i, f is either increasing with respect to the ith coordinate or decreasing with respect to the ith coordinate.

We prove Theorem 1 by means of the following much more general statement:

Proposition 1.

Let f1,ā€¦,f k :{Ā±1}nā†’{0,1}, be unate functions and let F:{Ā±1}nā†’{0,1} be defined as F(x)= āˆØ i = 1 k f i (x). Then

š”ø š•Š ( f ) = O n log ( k ) .

The application of Theorem 1 to machine learning is via a slightly different notion of noise sensitivity than that of the average sensitivity. In particular, we define the noise sensitivity as follows

Definition 2.

Let f:{Ā±1}nā†’{0,1} be a Boolean function. For a parameter Īµāˆˆ(0,1) we define the noise sensitivity of f with parameter Īµ to be

ā„• š•Š Īµ ( f ) : = Pr ( f ( x ) ā‰  f ( y ) )

where x and y are Bernoulli random variables where y is obtained from x by randomly and independently flipping the sign of each coordinate with probability Īµ.

Using this notation, we have that

Corollary 1.

If f:{Ā±1}nā†’{0,1} is the indicator function of the intersection of k halfspaces, and Īµāˆˆ(0,1) then

ā„• š•Š Īµ ( f ) = O Īµ log ( k ) .

Remark 1.

This is false in general for intersections of unate functions, since if f is the tribes function on n variables (which is unate) then ā„• š•Š Īµ (f)=Ī©(1) so long as Īµ=Ī©(logāˆ’1(n)).

Finally, using the L1 polynomial regression algorithm of [1], we obtain the following:

Corollary 2.

The concept class of intersections of k halfspaces with respect to the uniform distribution on {Ā±1}n is agnostically learnable with error o p t+Īµ in time n O ( log ( k ) Īµ āˆ’ 2 ) .

Proofs of the sensitivity bounds

The proof of Proposition 1 follows by a fairly natural generalization of one of the standard proofs for the case of a single unate function. In particular, if f:{Ā±1}nā†’{0,1} is unate, we may assume without loss of generality that f is increasing in each coordinate. In such a case, it is easy to show that

š”ø š•Š ( f ) = š”¼ f ( x ) āˆ‘ i = 1 n x i ā‰¤ š”¼ max 0 , āˆ‘ i = 1 n x i = O n .

In fact, this technique can be extended to prove bounds on the sensitivity of unate functions with given expectation. In particular, Lemma 1 below provides an appropriate bound. Our proof of Proposition 1 turns out to be a relatively straightforward generalization of this technique. In particular, we show that by adding the f i one at a time, the change in sensitivity is bounded by a similar function of the change in total expectation.

Lemma 1.

Let S:{Ā±1}nā†’{0,1} and let p=š”¼ S ( x ) , then

š”¼ S ( x ) āˆ‘ i = 1 n x i = O p n log ( 1 / p ) .

Proof.

Note that:

š”¼ S ( x ) āˆ‘ i = 1 n x i ā‰¤ āˆ« 0 āˆž Pr S ( x ) āˆ‘ i = 1 n x i > y dy ā‰¤ āˆ« 0 āˆž min p , Pr āˆ‘ i = 1 n x i > y dy ā‰¤ āˆ« 0 āˆž min p , exp āˆ’ Ī© y 2 / n dy ā‰¤ O āˆ« 0 āˆž min p , exp āˆ’ z 2 / n dz ā‰¤ O āˆ« 0 n log ( 1 / p ) pdz + āˆ« n log ( 1 / p ) āˆž exp āˆ’ z 2 / n dz ā‰¤ O ( p n log ( 1 / p ) ) .

We now prove Proposition 1.

Proof.

Let F m = āˆØ i = 1 m f i (x). Let S m (x)=F m (x)āˆ’Fmāˆ’1(x). Let p m =š”¼ S m ( x ) . Our main goal will be to show that š”øš•Š( F m )ā‰¤š”øš•Š( F m āˆ’ 1 )+O p m n log ( p m ) , from which our result follows easily.

Consider š”øš•Š( F m )āˆ’š”øš•Š( F m āˆ’ 1 ). We assume without loss of generality that f m is increasing in every coordinate.

š”ø š•Š ( F m ) āˆ’ š”ø š•Š ( F m āˆ’ 1 ) = āˆ‘ i = 1 n š”¼ F m ( x ) āˆ’ F m ( x i ) āˆ’ F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) ,

where xi denotes x with the ith coordinate flipped. We make the following claim:

Claim. For each x,i,

F m ( x ) āˆ’ F m ( x i ) āˆ’ F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) ā‰¤ x i F m ( x ) āˆ’ F m ( x i ) āˆ’ F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) .
(1)

Proof. Our proof is based on considering two different cases.

Case 1: f m (x)=f m (xi)=0

In this case, F m (x)=Fmāˆ’1(x) and F m (xi)=Fmāˆ’1(xi), and thus both sides of Equation 1 are 0.

Case 2: f m (x)=1 or f m (xi)=1

Note that replacing x by xi leaves both sizes of Equation 1 the same. We may therefore assume without loss of generality that x i =1. Since f m is increasing with respect to the ith coordinate, f m (x)ā‰„f m (xi). Since at least one of them is 1, f m (x)=1. Therefore, F m (x)=1. Therefore, since

x i F m ( x ) āˆ’ F m ( x i ) ā‰„ F m ( x ) āˆ’ F m ( x i ) ,

and

āˆ’ x i F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) ā‰„ āˆ’ F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) ,

Equation 1 follows.

By the claim we have that

š”ø š•Š ( F m ) āˆ’ š”ø š•Š ( F m āˆ’ 1 ) ā‰¤ āˆ‘ i = 1 n š”¼ x i F m ( x ) āˆ’ F m ( x i ) āˆ’ F m āˆ’ 1 ( x ) āˆ’ F m āˆ’ 1 ( x i ) = āˆ‘ i = 1 n š”¼ x i S m ( x ) āˆ’ S m ( x i ) = āˆ‘ i = 1 n š”¼ x i S m ( x ) āˆ’ āˆ‘ i = 1 n š”¼ ( āˆ’ y i ) S m ( y ) = 2 š”¼ S m ( x ) āˆ‘ i = 1 n x i = O ( p m n log ( 1 / p m ) ) .

Where the on the third line above, we are letting y=xi, and the last line is by Lemma 1.

Hence, we have that

š”ø š•Š ( F ) = āˆ‘ m = 1 k š”ø š•Š ( F m ) āˆ’ š”ø š•Š ( F m āˆ’ 1 ) = O n āˆ‘ m = 1 k p m log ( 1 / p m ) .

Let P=š”¼[F(x)]= āˆ‘ m = 1 k p m . By concavity of the function x log ( 1 / x ) for xāˆˆ(0,1), we have that

š”ø š•Š ( F ) = O n P log ( k / P ) = O n log ( k ) .

This completes our proof.

Theorem 1 follows from Proposition 1 upon noting that 1āˆ’f is a disjunction of k linear threshold functions, each of which is unate. Our proof of Theorem 1 shows that the bound can be tight only if a large number of the halfspaces cut off an incremental volume of roughly 1/k. It turns out that this bound can be achieved when we take a random collection of halfspaces with such volumes. Before we begin to prove Theorem 2, we need the following Lemma:

Lemma 2.

For an integer n and 1/2>Īµ>2āˆ’n there exists a linear threshold function f:{Ā±1}nā†’{0,1} so that

š”¼ x f ( x ) ā‰„ Īµ ,

and

š”ø š•Š ( f ) = Ī© š”¼ x [ f ( x ) ] n log ( 1 / Īµ ) .

Proof.

This is easily seen to be the case if we let f(x) be the indicator function of āˆ‘ i = 1 n x i >t for t as large as possible so that this event takes place with probability at least Īµ.

Proof.

We note that it suffices to show that there is such as f given as the indicator function of a union of at most k half-spaces, as 1āˆ’f will have the same average sensitivity and will be the indicator function of an intersection. Let Īµ=1/k, and let f be the function given to us in Lemma 2. We note that if š”¼[f(x)]>1/4, then f is sufficient and we are done. Otherwise let m=āŒŠ1/ 4 š”¼ [ f ( x ) ] āŒ‹ā‰¤k. For sāˆˆ{Ā±1}n let f s (x)=f(s1x1,ā€¦,s n x n ). We note for each s that f s (x) is a linear threshold function with š”¼[ f s (x)]=š”¼[f(x)] and š”øš•Š( f s )=š”øš•Š(f).

Let

F ( x ) = āˆØ i = 1 m f s i ( x )

for s i independent random elements of {Ā±1}n. We note that F(x) is always the indicator of a union of at most k half-spaces, but we also claim that

š”¼ s i [ š”ø š•Š ( F ) ] = Ī© n log ( k ) .

This would imply our result for appropriately chosen values of the s i .

We note that š”øš•Š(F) is 21āˆ’n times the number of pairs of adjacent elements x,y of the hypercube so that F(x)=1,F(y)=0. This in turn is at least 21āˆ’n times the sum over 1ā‰¤iā‰¤m of the number of pairs of adjacent elements of the hypercube x,y so that f s i (x)=1, f s i (y)=0 and so that f s j (x)= f s j (y)=0 for all jā‰ i.

On the other hand, for each i, 21āˆ’n times the number of pairs of adjacent elements x,y so that f s i (x)=1, f s i (y)=0 is

š”ø š•Š ( f s i ) = š”ø š•Š ( f ) = Ī© š”¼ [ f ( x ) ] n log ( k ) = Ī© m āˆ’ 1 n log ( k ) .

For each of these pairs, we consider the probability over the choice of s j that f s j (x)=1 or f s j (y)=1 for some jā‰ i. We note that for each fixed x and j that

Pr s j f s j ( x ) = 1 = š”¼ s j f s j ( x ) = š”¼ s j f x ( s j ) = š”¼ z f ( z ) ā‰¤ 1 4 m .

Thus, by the union bound, the probability that either f s j (x)=1 or f s j (y)=1 for some jā‰ i is at most 1/2. Therefore, the expected number of adjacent pairs x,y with f s i (x)=1, f s i (y)=0 and f s j (x)= f s j (y)=0 for all jā‰ i is at least š”øš•Š( f s j )/2. Therefore,

š”¼ s i š”ø š•Š ( F ) ā‰„ āˆ‘ i = 1 m š”ø š•Š ( f ) / 2 = mĪ© m āˆ’ 1 n log ( k ) = Ī© n log ( k ) ,

as desired. This completes our proof.

Learning theory application

The proofs of Corollaries 1 and 2 are by what are now fairly standard techniques, but are included here for completeness. The proof of Corollary 1 is by a technique of Diakonikolas et al. in [8] for bounding the noise sensitivity in terms of the average sensitivity.

Proof.

As the noise sensitivity is an increasing function of Īµ for Īµ<1/2, we may round Īµ down to 1/āŒˆĪµāˆ’1āŒ‰, and thus it suffices to consider Īµ=1/m for some integer m. We note that the pair of random variables x, y used to define the noise sensitivity with parameter Īµ can be generated in the following way:

  1. 1.

    Randomly divide the n coordinates into m bins.

  2. 2.

    Randomly assign each coordinate a value in {Ā±1} to obtain z.

  3. 3.

    For each bin randomly pick b i āˆˆ{Ā±1}. Obtain x from z by multiplying all coordinates in the i th bin by b i for each i.

  4. 4.

    Obtain y from x by flipping the sign of all coordinates in a randomly chosen bin.

We note that this produces the same distribution on x and y since x is clearly a uniform element of {Ā±1}n and the ith coordinate of y differs from the corresponding coordinate of x if and only if i lies in the bin selected in step 4. This happens independently and with probability 1/m for each coordinate.

Next let f be the indicator function of an intersection of at most k halfspaces. Note that after the bins and z are picked in steps 1 and 2 above that f(x) is given by g(b) where g is the indicator function of an intersection of at most k halfspaces in m variables. In the same notation, f(y)=g(bā€²) where bā€² is obtained from b by flipping the sign of a single random coordinate. Thus, by definition, Pr g ( b ) ā‰  g ( b ā€² ) = 1 m š”øš•Š(g). Hence,

ā„• š•Š Īµ ( f ) = š”¼ g š”ø š•Š ( g ) m ā‰¤ O log ( k ) m m = log ( k ) m = Īµ log ( k ) .

This completes our proof.

Corollary 2 will now follow by using this bound to bound the weight of the higher degree Fourier coefficients of such an f and then using the L1 polynomial regression algorithm of [1].

Proof.

Let f be the indicator function of an intersection of k halfspaces. Let f have Fourier transform given by

f = āˆ‘ S āŠ‚ [ n ] Ļ‡ S f Ģ‚ ( S ) .

It is well known that for Ļāˆˆ(0,1) that

ā„• š•Š Ļ ( f ) = 2 āˆ‘ S āŠ‚ [ n ] 1 āˆ’ ( 1 āˆ’ 2 Ļ ) | S | f Ģ‚ ( S ) 2 .

Therefore, we have that

ā„• š•Š Ļ ( f ) ā‰« āˆ‘ | S | > 1 / Ļ f Ģ‚ ( S ) 2 .

By Corollary 1, this tells us that

āˆ‘ | S | > 1 / Ļ f Ģ‚ ( S ) 2 = O Ļ log ( k ) .

Setting Ļ=Īµ2/(C log(k)) for sufficiently large values of C yields

āˆ‘ | S | > C log ( k ) Īµ āˆ’ 2 f Ģ‚ ( S ) 2 < Īµ.

Our claim now follows from [1] Remark 4.