<?xml-model href='http://www.tei-c.org/release/xml/tei/custom/schema/relaxng/tei_all.rng' schematypens='http://relaxng.org/ns/structure/1.0'?><TEI xmlns="http://www.tei-c.org/ns/1.0">
	<teiHeader>
		<fileDesc>
			<titleStmt><title level='a'>Approximately Hadamard Matrices and Riesz Bases in Random Frames</title></titleStmt>
			<publicationStmt>
				<publisher></publisher>
				<date>2023 Spring</date>
			</publicationStmt>
			<sourceDesc>
				<bibl> 
					<idno type="par_id">10432841</idno>
					<idno type="doi"></idno>
					<title level='j'>International mathematics research notices</title>
<idno>1073-7928</idno>
<biblScope unit="volume"></biblScope>
<biblScope unit="issue"></biblScope>					

					<author>Xiaoyu Dong and Mark Rudelson</author>
				</bibl>
			</sourceDesc>
		</fileDesc>
		<profileDesc>
			<abstract><ab><![CDATA[An n × n matrix with ±1 entries which acts on R n as a scaled isometry is called Hadamard. Such matrices exist in some, but not all dimensions. Combining number-theoretic and probabilistic tools we construct matrices with ±1 entries which act as approximate scaled isometries in R n for all n ∈ N. More precisely, the matrices we construct have condition numbers bounded by a constant independent of n.Using this construction, we establish a phase transition for the probability that a random frame contains a Riesz basis. Namely, we show that a random frame in R n formed by N vectors with independent identically distributed coordinate having a non-degenerate symmetric distribution contains many Riesz bases with high probability provided that N ≥ exp(Cn). On the other hand, we prove that if the entries are subgaussian, then a random frame fails to contain a Riesz basis with probability close to 1 whenever N ≤ exp(cn), where c < C are constants depending on the distribution of the entries.]]></ab></abstract>
		</profileDesc>
	</teiHeader>
	<text><body xmlns="http://www.tei-c.org/ns/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction and main results</head><p>Let n &lt; N be natural numbers. A set of vectors X 1 , . . . , X N &#8712; R n is called a frame if</p><p>for all x &#8712; R n . Here R &#8805; 1 is an absolute constant called the frame constant, and K(n, N ) &gt; 0 is some function of n and N . The notation &#8741;x&#8741; 2 stands for the Euclidean norm of the vector x = (x 1 , . . . , x n ):</p><p>In the last 40 years, frame theory became a well-developed area of applied mathematics, see <ref type="bibr">[4]</ref>, <ref type="bibr">[5]</ref>, <ref type="bibr">[6]</ref>, and the references therein. A frame can intuitively be regarded as overcomplete basis in R n . Because of this property, frames became a valuable tool in signal transmission. A signal which is viewed as an n-dimensional vector can be encoded by the sequence of its inner products with the frame vectors. If this sequence is transmitted over a communication line, then the original signal can be reconstructed even if part of the coefficients is lost or corrupted in the process of transmission. Moreover, this encoding is robust, which means that if the inner products are evaluated with some noise, then the reconstructed version will be close to the original one with the error depending on the noise magnitude.</p><p>One of the most popular classes of frames in algorithmic applications is the set of random frames. Such frames became also the method of choice in compressed sensing where one needs to reconstruct a low complexity signal from a small number of linear measurements, see, e.g., <ref type="bibr">[21]</ref>. For example, if complexity is measured as the size of the support, and the support itself is unknown, the random frames provide robust recovery with optimal or almost optimal theoretical guarantees.</p><p>To construct a random frame, consider a random vector X &#8712; R n with centered uncorrelated coordinates of unit variance. In other words, assume that E X = 0 and E XX &#8868; = I n . Let vectors X 1 , . . . , X N be independent copies of X. The Law of Large Numbers implies that lim</p><p>and thus, with probability close to 1,</p><p>for all x &#8712; R n provided that N = N (&#949;) is sufficiently large. Another, trivial way to construct a frame is to take several bases in R n and concatenate them. This allows an exact reconstruction of the transmitted signal if the number of corrupted coordinates is relatively small. Indeed, one can reconstruct the original vector from the set of transmitted coordinates for each basis separately and keep the copy which is repeated many times. While in practice the random frames perform better than such concatenated bases, it leads to a question whether a random frame contains a copy or copies of a nice basis. More precisely, a sequence of n vectors v 1 , . . . , v n &#8712; R n is called a Riesz basis if it possesses the frame property <ref type="bibr">(1.1)</ref>. This property ensures that the reconstruction is robust, i.e., that the reconstructed vector is close to the original one if the coordinates are distorted by adding a small noise. These considerations lead to a natural question of determining the values of N for which a random frame {X 1 , . . . , X N } &#8834; R n contains one or many Riesz bases with high probability.</p><p>This problem can be conveniently translated to the language of random matrices. Namely, for an n &#215; N matrix A, define its singular values as</p><p>where s j (A) = &#955; j (AA &#8868; ), and &#955; 1 (AA &#8868; ), . . . , &#955; n (AA &#8868; ) are eigenvalues of AA &#8868; arranged in the decreasing order. Also, define the condition number of A as</p><p>using the convention that &#954;(A) = &#8734; whenever s min (A) = 0. With this notation, the frame property (1.1) can be rewritten as &#954;(A n,N ) &#8804; C where A n,N is the n &#215; N matrix with columns X 1 , . . . , X N . Thus, the problem of existence of a Riesz basis in a random frame can be recast as the question of existence of one or many well-conditioned square n &#215; n submatrices of an n &#215; N random matrix A n,N with i.i.d. entries. Our first main result shows that the probability of finding such a submatrix undergoes a phase transition when N is exponential in terms of n. Since the upper and the lower bound hold under somewhat different assumptions, we formulate them separately.</p><p>Denote by [N ] the set {1, . . . , N }. Let A be an n &#215; N matrix. If I &#8834; [N ], denote by A I the submatix of A whose columns belong to I. The following theorem shows that if N is exponential in n, then with high probability, the n &#215; N random matrix has many square submatrices with uniformly bounded condition numbers. In the language of frames, it means that a random frame with exponentially many vectors contains a large number of Riesz bases whose frame constants are uniformly bounded.</p><p>Theorem 1.1. Let A be an n&#215;N matrix with i.i.d. symmetric non-degenerate entries. Then there exist constants c, C, &#945;, &#946; &gt; 0 depending on the distribution of entries of A with the following property. Assume that N &#8805; exp(Cn).</p><p>Then there exists L &#8805; exp(cn)</p><p>The strategy of proving Theorem 1.1 relies on using a certain deterministic n &#215; n matrix V having a bounded condition number. Denote by Col j (M ) the j-th column of the matrix M . We partition the set of integers [N ] into n subsets I 1 , . . . , I n of approximately the same size and show that with high probability, the set {Col i (A)} i&#8712;I j contains many columns close to Col j (V ). Condition on the event that such columns exist and form an n &#215; n matrix B taking one column from each set. Then conditionally this matrix can be viewed as a noisy version of the matrix V . This allows to show that with high probability, the matrix B has a bounded condition number as well.</p><p>The key to this strategy is a successful choice of the pattern matrix V . Since we strive to prove Theorem 1.1 under minimal assumptions, the choice of V becomes a non-trivial task. Indeed, the requirement that a column of A can be close to a column of V with a non-negligible probability forces us to look for matrices V whose entries are in the support of the distribution of an entry of A. The latter can be as small as two points, a and -a because of the symmetry assumption. Thus, we need to to construct V as a scaled copy of a matrix with &#177;1 entries. Such matrices are known in some cases. For instance, the condition number of any Hadamard matrix is one. An n &#215; n matrix H is called Hadamard if n -1/2 H is an isometry. The earliest result on the existence of Hadamard matrices was probably proved by Sylvester <ref type="bibr">[20]</ref> who showed that Hadamard matrices exist for dimension 2 k where k is any nonnegative integer (these matrices are now called Wash since their rows are Walsh functions). Hadamard matrices is a well-studied subject, and a number of constructions of such matrices are available, see e.g., the books <ref type="bibr">[1]</ref>, <ref type="bibr">[12]</ref>, and the references therein. In particular, Wallis <ref type="bibr">[23]</ref> proved that if p &gt; 3 is an integer, then there exists an Hadamard matrix of order 2 t p, where t = &#8970;2log 2 (p -3)&#8971;. Craigen <ref type="bibr">[7]</ref> improved Wallis's result by showing that for any odd number p, there exists an Hadamard matrix of order 2 t p, where t = 4 1  6 log 2 ((p -1)/2) + 2. Recently, de Launey <ref type="bibr">[10]</ref> studied the asymptotic existence of Hadamard matrices and concluded that for any &#1013; &gt; 0, the set of odd numbers k for which there is a Hadamard matrix of order k2 2+[&#1013;log 2 (k)] always has positive density in the set of natural numbers. Yet, the dimensions in which Hadamard matrices were constructed are rare.</p><p>This leads us to a task of constructing approximately Hadamard matrices, i.e., matrices with &#177;1 entries and bounded condition numbers. Some constructions of matrices with properties simlar to Hadamard's are available. For example, Banica, Nechita, and &#379;yczkowski <ref type="bibr">[2]</ref> defined an almost Hadamard matrix to be a N dimensional real square matrix H, such that H/ &#8730; N is orthogonal, and is a local maximum of the &#8467; 1 -norm of the entries on the orthogonal group O(N ). They showed the existence of almost Hadamard matrices under some special assumptions. There is also a notion of quasi-Hadamard matrix <ref type="bibr">[2,</ref><ref type="bibr">15]</ref>, which is defined as a square matrix with {-1, 1} entries that maximizes the absolute value of the determinant, but there are only very limited results on the existence of those matrices. In summary, no existing construction is directly related to our purpose.</p><p>The second main result of the paper is the following theorem asserting the existence of an approximately Hadamard matrix in all dimensions.</p><p>Theorem 1.2. There exist constants 0 &lt; c &lt; C such that for any n &#8712; N, one can find an n &#215; n matrix V with &#177;1 entries satisfying</p><p>The proof of Theorem 1.2 relies on Vinogradov's theorem from analytic number theory and combines number-theoretic and probabilistic ideas. The details are presented in Section 2. After Theorem 1.2 is proved, we prove Theorem 1.1 in Section 3.</p><p>The conclusion of Theorem 1.1 holds under minimal assumptions on the distribution of entries. If we assume that the entries of the matrix are subgaussian, then the bound of Theorem 1.1 becomes sharp. Recall that a random variable X is called subgaussian if there is a &gt; 0 such that</p><p>If X is subgaussian then the smallest number a having this property is called the subgaussian norm of X and denoted &#8741;X&#8741; &#968; 2 . Subgaussian random variables form a large family containing many naturally arising ones, see, e.g. <ref type="bibr">[22]</ref>.</p><p>The next theorem shows that finding a submatrix with a bounded condition number requires an exponential number of columns for matrices with subgaussian entries.</p><p>Theorem 1.3. Let X be a centered subgaussian random variable. Then there exist C, c, c, t 0 &gt; 0 depending only on</p><p>with the following property. Let t &gt; t 0 , and assume that N &#8804; exp c t 4 n . Let A be an n &#215; N matrix whose entries are independent copies of X. Then</p><p>We prove Theorem 1.3 in Section 4. Its proof is easier than that of Theorem 1.1 and relies on the Hanson-Wright inequality <ref type="bibr">[19]</ref>.</p><p>Acknowledgment. The second author is grateful to Marcin Bownik for helpful discussions and bringing his attention to the problem. Part of this work was done when the second author visited the Weizmann Institute of Science. He is grateful to the Institute for its hospitality. The authors thank a referee for thoroughly checking the manuscript and correcting many typos.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Approximately Hadamard matrices</head><p>In this section we construct an n &#215; n matrix with &#177;1 entries whose scaled copy acts on R n as an approximate isometry. More precisely, for any sufficiently large n, we construct an n &#215; n matrix V such that its condition number &#954;(V ) is bounded by an absolute constant.</p><p>We use standard matrix norms below. Namely, &#8741;A&#8741; stands for the operator norm of an n &#215; m matrix A = (a i,j ), and &#8741;A&#8741; HS stands for its Hilbert-Schmidt or Frobenius norm: We will apply an above mentioned result of Wallis <ref type="bibr">[23]</ref> showing that Hadamard matrices exist in dimensions close to n 3 . Lemma 2.1. There is l 0 &#8712; N such that for any l &gt; l 0 , there exists an Hadamard matrix of dimension m(l) with m(l) = 2 2&#8968;log 2 (l-3)&#8969; l.</p><p>We will need the following corollary.</p><p>Corollary 2.2. For any &#949; &gt; 0, there exists N (&#949;) such that for any n &gt; N 0 (&#949;), one can find an even number m &#8712; [(1 -&#949;)n, (1 + &#949;)n] for which there exists an Hadamard matrix of size m &#215; m.</p><p>Proof. For any n &gt; 12, there exists a unique k</p><p>. By Lemma 2.1, there exists an Hadamard matrix of size m &#215; m. Since</p><p>Since the tensor product of Hadamard matrices is an Hadamard matrix, and there are Hadamard matrices of sizes 2 &#215; 2 and 4 &#215; 4, there exist Hadamard matrices of sizes 2m(l) and 4m(l) for all l &gt; l 0 . This allows completing the proof of the corollary in the remaining cases when 2 2k+1 (2</p><p>The aim of this section is to construct approximately Hadamard matrices in any dimension, i.e. matrices whose condition number is O(1). To this end, we use a construction of approximately Hadamard matrices of a prime size.</p><p>Let q &#8712; N be an odd prime number. For k &#8712; Z q , denote e q (k) = exp 2&#960;i k q .</p><p>Define the Fourier transform on Z q setting v(j) = k&#8712;Zq v(k)e q (jk) for a vector v &#8712; C Zq and j &#8712; Z q .</p><p>Lemma 2.3. Let q be an odd prime number. Then there exists a vector u q &#8712; {-1, 1} Zq such that |&#251; q (j)| -&#8730; q &#8804; &#8730; q&#948; q for any j &#8712; Z q with &#948; q = Cq -1/4 &#8730; log q.</p><p>Proof. The construction closely follows the one in [14, Proposition 3.2], which in turn originates in <ref type="bibr">[16,</ref><ref type="bibr">Theorem 9.2]</ref>. Let v : Z q &#8594; {-1, 1} be the Legendre symbol (quadratic character mod q). More precisely, let Q = {k &#8712; Z q : k = j 2 (mod q) for some j &#8712; Z q } \ {0} be the set of quadratic residues, and set</p><p>Then by a standard result on the Gauss sum, see e.g., [13, Proposition 6.3.2.; p.71], we have</p><p>The difference between v and the desired function u q is that v(0) = 0 and v(0) = 0. We will perturb v replacing some of its coordinates by -1 to change the value of v(0) as required while keeping the other Fourier coefficients close to their original values. To this end, consider a sequence of i.i.d. random variables {X k } k&#8712;Q such that</p><p>Then u q : Z q &#8594; {-1, 1}, so we only have to check the values of the Fourier coefficients. Let us start with the expectations. We have</p><p>and </p><p>where the last equality follows if we fix k + l and sum over k -l first. Thus, |E &#251;q (j) -v(j)| &#8804; 2 + q -1/2 for any j &#8712; Z q \ {0}, and so | E &#251;q (j)| -&#8730; q &#8804; 3 for all j &#8712; Z q .</p><p>The quantity &#251;q (j)-E &#251;q (j) is a linear combination of i.i.d. centered random variables X k -E X k , k &#8712; Q with coefficients e q (jk) whose absolute value is bounded by 1. Therefore, Bernstein's inequality yields P (| &#251;q (j) -E &#251;q (j)| &gt; t) &#8804; 2 exp -c min(t 2 q -1/2 , t) for all t &gt; 0 and j &#8712; Z q . Setting t = Cq 1/4 &#8730; log q and taking the union bound over j &#8712; Z q , we obtain P | &#251;q (j) -E &#251;q (j)| &#8804; Cq 1/4 log q for all j &#8712; Z q &#8805; 1 -q -1 &gt; 0 if the constant C &gt; 0 is chosen sufficiently large. The lemma follows. &#9633; Corollary 2.4. Let q be an odd prime number. There exists a q &#215; q matrix U q with &#177;1 entries such that &#8730; q(1 -&#948; q ) &#8804; s min (U q ) &#8804; s max (U q ) &#8804; &#8730; q(1 + &#948; q )</p><p>with &#948; q = Cq -1/4 &#8730; log q.</p><p>Proof. Represent Z q as {1, . . . , q}, and let u q : {1, . . . , q} &#8594; {-1, 1} be the vector defined in Lemma 2.3. Let U q be the circulant matrix with the first row u q . A circulant matrix is diagonal in the Fourier basis, see e.g., [9, Theorem 3.2.1; p. 72]. Therefore, the singular values of U q are the absolute values of its eigenvalues which are the Fourier coefficients of the generating vector u q . The result follows from Lemma 2.3. &#9633; Remark 2.5. Corollary 2.4 implies that the matrix U q satisfies U q (U q ) &#8868; -qI q &#8804; 3&#948; q q. (2.1) Inequality (2.1) will be used later in the proof of Theorem 1.2.</p><p>With this auxiliary construction in place, we can prove the main result of this section, namely Theorem 1.2.</p><p>Proof of Theorem 1.2. The proof of this theorem combines a deterministic construction of number-theoretic nature with a probabilistic argument. Without loss of generality, we can assume that n is larger than some number n 0 chosen in advance. Indeed, after the statement of the theorem is proved for n &#8805; n 0 , we can adjust the constants c and C appropriately to make it hold for all n &#8712; N.</p><p>We start with the case when n is even. Let &#949; &gt; 0 be a number to be chosen later. A combination of the Prime Number Theorem and Vinogradov's sum of three primes theorem <ref type="bibr">[18]</ref>, yields that there exists N = N (&#949;) such that any even n &gt; N has a decomposition (2.2)</p><p>where q 1 , . . . , q 4 are prime numbers. Indeed, by the Prime Number Theorem, there exists a prime number q 1 such that (1 -&#949;/2) n 4 &#8804; q 1 &#8804; (1 + &#949;/2) n 4 . Then m = n -q 1 is odd, and thus by a stronger version of Vinogradov's theorem, it can be decomposed as</p><p>and q 2 , q 3 , q 4 are primes. This immediately implies (2.2). Actually, decompositions with bounds tighter than (2.3) are available. More precisely, one can find a representations such as (2.3) with |q j -m/3| &lt; m &#952; for some &#952; &#8712; (0, 1), see e.g., <ref type="bibr">[3,</ref><ref type="bibr">11,</ref><ref type="bibr">17]</ref>. However, the weaker version presented above will be sufficient for our purposes. Without loss of generality, assume that q 1 &#8805; &#8226; &#8226; &#8226; &#8805; q 4 =: q. We will consider the case q 3 &gt; q 4 first. This is the most non-trivial case, and the other ones will be treated in the same way after obvious modifications. For j = {1, . . . , 4}, let U j be the matrix U q j constructed in Corollary 2.4, and denote by U top j the submatrix formed by the q top rows of U j . For j &#8712; {1, 2, 3}, denote by U bottom j the submatrix of U j formed by its q j -q bottom rows. We will construct the matrix W = V &#8868; in the following block form:</p><p>where the matrix W top consists of the upper 4 block rows of W , and W bottom consists of the lower three. Here W j,k is a q &#215; q k matrix if 1 &#8804; j, k &#8804; 4 and a (q j-4 -q) &#215; q k matrix if j = 5, 6, 7, 1 &#8804; k &#8804; 4.</p><p>Let us define the matrices W j,k . The matrix W top will be deterministic, and the matrix W bottom will consist of deterministic and random blocks. For 1 &#8804; j, k &#8804; 4, set W j,k = &#949; j,k U top k , where &#949; j,k , j, k &#8712; {1, . . . , 4} form a 4 &#215; 4 Walsh matrix:</p><p>Now, let us define the matrices W j,k for j = 5, 6, 7. Set W j,j-4 = U bottom j-4 . For j = 5, 6, 7 and k &#824; = j -4, let W j,k be a random matrix with i.i.d. Rademacher entries.</p><p>Consider the (4q) &#215; (4q) matrix W top (W top ) &#8868; first. The diagonal blocks of this matrix are close to nI q . More precisely, for any j &#8712; {1, . . . , 4},</p><p>where the first inequality follows since</p><p>&#8868; and the second one from (2.1).</p><p>Let us consider the off-diagonal blocks now. If i &#824; = j, i, j &#8712; {1, . . . , 4} then similarly</p><p>where the first estimate follows from the triangle inequality, and the second one from</p><p>Combining the two inequalities, we obtain</p><p>for all sufficiently large n.</p><p>Let us introduce auxiliary (n -4q) &#215; n matrices S and R defined by</p><p>In other words, R is the random part of the matrix W bottom , i.e., Recall that n -4q &#8804; 4&#949;n. In view of Corollary 2.4 and inequality (2.1),</p><p>Also,</p><p>Let R be an (n -4q) &#215; n matrix with i.i.d. Rademacher entries. Then a simple symmetrization argument yields</p><p>which in combination with [19, Theorem 3.2] implies that</p><p>Another application of symmetrization yields</p><p>Fix a matrix R for which</p><p>at the same time. Let x &#8712; S n-1 . Following the previous convention, we write</p><p>where x top &#8712; R 4q and x bottom &#8712; R n-4q . Assume first that x bottom 2 &#8805; &#951;, where the constant &#951; &gt; 0 will be chosen below. Then</p><p>Assume now that &#8741;x bottom &#8741; 2 &lt; &#951;. Then &#8741;x top &#8741; 2 &gt; 1 -&#951;, and (2.4) yields</p><p>Choosing the parameters &#949; and &#951; sufficiently small, we can reconcile the two restrictions, i.e., select &#949;, &#951; so the inequalities</p><p>(1 -2&#949;)&#951; -C 1 &#8730; &#949; &gt; &#951;/2 and C 2 &#951; &lt; 1 4 hold at the same time. With this choice, the previous argument shows that</p><p>&#8730; n for all x &#8712; S n-1 , which means that</p><p>Obtaining a bound for s max (W ) is easier. Inequalities (2.4) and (2.5) imply</p><p>This in combination with (2.6) yields</p><p>which proves the theorem in the case q 3 &gt; q.</p><p>If q j = q for some j &#8712; {1, 2, 3}, then we repeat the same argument with the block rows of W containing q j -q = 0 rows removed. For instance, if q 1 &gt; q and q 2 = q 3 = q, then we consider the n &#215; n matrix W with W top being the same as in the previous case and W bottom = (W 5,1 &#8226; &#8226; &#8226; W 5,4 ) which is a (q 1 -q) &#215; n matrix.</p><p>This completes the proof of the theorem in the case when n is even.</p><p>Assume that n is odd. By Corollary 2.2, if n is sufficiently large, then we can find an even number</p><p>n 4 for which there exists an Hadamard matrix V of size m &#215; m. Note that n -m is odd and</p><p>so using Vinogradov's theorem again, we obtain a decomposition n -m = q 1 + q 2 + q 3 , where q 1 , q 2 , q 3 are prime numbers and 1-&#949; 4 n &#8804; q j &#8804; 1+&#949; 4 n. At this point we can apply the same argument we used in the case of an even n with one of the matrices U 1 , . . . , U 4 replaced by V . This completes the proof of the theorem. &#9633;</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Submatrices with a small condition number</head><p>In this section we prove Theorem 1.1. As was explained in the Introduction, the proof relies on finding columns of A which are close to columns of a scaled copy of the V constructed in the previous section. Conditioned on the event that such selection is possible, we prove that with high probability, the constructed submatrix has a bounded condition number. We start with the latter task, namely with analyzing a random matrix close to V . Lemma 3.1. Let V be an n &#215; n matrix with &#177;1 entries such that</p><p>There exists &#948; &#8712; (0, 1) for which any n &#215; n matrix Y with i.i.d. entries Y i,j such that E Y i,j = 0 and |Y i,j | &#8804; &#948; a.s. satisfies</p><p>Proof. The proof of Lemma 3.1 uses the basic net argument, see e.g., <ref type="bibr">[22]</ref>. Since Y has i.i.d. centered subgaussian entries with &#8741;Y i,j &#8741; &#968; 2 &#8804; &#8741;Y i,j &#8741; &#8734; &#8804; &#948;,</p><p>Therefore,</p><p>as we can always assume that C 3.1 &#8805; 1 and choose &#948; sufficiently small. Similarly,</p><p>where as before, the last inequality holds for any sufficiently small &#948;. The result follows by combining the two bounds above. &#9633;</p><p>We now proceed to proving the main result, Theorem 1.1.</p><p>Proof ot Theorem 1.1. Since the distribution of entries of A is non-degenerate, there exists a &gt; 0 such that for any &#957; &gt; 0</p><p>By the symmetry of distribution, we also have the same property for -a in place of a.</p><p>Let &#948; &gt; 0 be as in Lemma 3.1, and denote &#957; = a 4 &#948;. Let Z be a random variable having the same distribution as a i,j conditioned on the event that |a i,j -a| &#8804; &#957;. More precisely, for a Borel set E &#8834; R, set</p><p>Then R is a centered random variable such that</p><p>Let C &gt; 0 be a constant to be chosen later, and assume that N &#8805; exp(2Cn). Partition [N ] into a union of sets I 1 , . . . , I n such that</p><p>Let V be the n&#215;n matrix with &#177;1 entries constructed in Theorem 1.2. Denote its columns by V 1 , . . . , V n and the columns of A by A 1 , . . . , A N . Let M be a number to be chosen later. Let E be the event that for any j &#8712; [n], there exist at least M numbers k &#8712; I j with</p><p>If E occurs, denote by k(j, 1), . . . , k(j, M ) the first M numbers k &#8712; I j having this property. Then conditioned on E, for any m &#8712; [M ], the matrix A E,m with columns A k(1,m) , . . . , A k(n,m) has the same distribution as E Z &#8226; (V + Y ), where Y is an n &#215; n random matrix whose entries have the form Y i,j = V i,j R i,j , where R i,j are independent copies of R. In view of Lemma 3.1, this implies that for any m &#8712; [M ],</p><p>Since conditionally on E, the matrices A E,1 , . . . , A E,M are independent, Bernstein's inequality allows to conclude that</p><p>To complete the proof, we have to show that the probability of E c is small. To this end, denote &#951; = P(|a i,j -a| &#8804; &#957;). Then by the symmetry of distribution of the entries of A,</p><p>where the last inequality holds if C = C(&#951;) is chosen sufficiently large. Note that the events &#8741;A k -aV j &#8741; &#8734; &#8804; &#957; are independent for all k &#8712; I j . At this point, Bernstein's inequality yields</p><p>Therefore,</p><p>Let L = M/2, and &#945; = 4 C 3.1 c 3.1 . Combining the previous inequalities, we obtain that</p><p>for an appropriate &#946; &gt; 0. This completes the proof of the theorem. &#9633;</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4.</head><p>No submatrices with a small condition number</p><p>In this section, we prove Theorem 1.3.</p><p>Proof. Without loss of generality, we can assume that &#8741;X&#8741; 2 = (E X 2 ) 1/2 = 1. Throughout the proof, we denote by C, c, c &#8242; , etc. constants depending only on &#8741;X&#8741; &#968; 2 . Consider an n &#215; n random matrix B whose entries are independent copies of X. We claim that  </p></div></body>
		</text>
</TEI>
