<?xml-model href='http://www.tei-c.org/release/xml/tei/custom/schema/relaxng/tei_all.rng' schematypens='http://relaxng.org/ns/structure/1.0'?><TEI xmlns="http://www.tei-c.org/ns/1.0">
	<teiHeader>
		<fileDesc>
			<titleStmt><title level='a'>The Rate of Interactive Codes Is Bounded Away from 1</title></titleStmt>
			<publicationStmt>
				<publisher>ACM</publisher>
				<date>01/01/2023</date>
			</publicationStmt>
			<sourceDesc>
				<bibl> 
					<idno type="par_id">10488499</idno>
					<idno type="doi"></idno>
					<title level='j'>Conference proceedings of the annual ACM Symposium on Theory of Computing</title>
<idno>0734-9025</idno>
<biblScope unit="volume"></biblScope>
<biblScope unit="issue"></biblScope>					

					<author>Klim Efremenko</author><author>Gillat Kol</author><author>Dmitry Paramonov</author><author>Raghuvansh R. Saxena</author>
				</bibl>
			</sourceDesc>
		</fileDesc>
		<profileDesc>
			<abstract><ab><![CDATA[Kol and Raz [STOC 2013]  showed how to simulate any alternating two-party communication protocol designed to work over the noiseless channel, by a protocol that works over a stochastic channel that corrupts each sent symbol with probability 𝜖 > 0 independently, with only a 1 + O √︁ H(𝜖) blowup to the communication. In particular, this implies that the maximum rate of such interactive codes approaches 1 as 𝜖 goes to 0, as is also the case for the maximum rate of classical error correcting codes. Over the past decade, followup works have strengthened and generalized this result to other noisy channels, stressing on how fast the rate approaches 1 as 𝜖 goes to 0, but retaining the assumption that the noiseless protocol is alternating.In this paper we consider the general case, where the noiseless protocols can have arbitrary orders of speaking. In contrast to Kol-Raz and to the followup results in this model, we show that the maximum rate of interactive codes that encode general protocols is upper bounded by a universal constant strictly smaller than 1. To put it differently, we show that there is an inherent blowup in communication when protocols with arbitrary orders of speaking are faced with any constant fraction of errors 𝜖 > 0. We mention that our result assumes a large alphabet set and resolves the (nonbinary variant) of a conjecture by Haeupler [FOCS 2014].]]></ab></abstract>
		</profileDesc>
	</teiHeader>
	<text><body xmlns="http://www.tei-c.org/ns/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>One of the gems in Shannon's famous 1948 paper introducing information theory <ref type="bibr">[17]</ref> is the channel capacity formula, that gives the maximum rate possible for an error correcting code over any discrete memoryless channel. Recall that an error correcting code with rate &#119903; allows one party to reliably communicate a message consisting of &#119899; symbols to a remote second party, with a negligible probability of error, by sending only &#119899; &#8226; (1/&#119903; + &#119900; (1)) symbols over the channel. Let C &#915;,&#120598; be the symmetric channel with alphabet set &#915; and noise rate &#120598;. <ref type="foot">1</ref> The channel capacity formula shows that the maximum rate over C &#915;,&#120598; approaches 1 as the noise rate &#120598; approaches 0. For instance, the maximum rate over C {0,1},&#120598; is 1 -H(&#120598;), where H is the binary entropy function.</p><p>Schulman's groundbreaking work <ref type="bibr">[15]</ref> studied error correcting codes in the &#322;two-way&#382; setting, where there are noisy channels between the two communicating parties in both directions. Such error correcting codes are called interactive codes and they allow the encoding of interactive protocols, which may consist of many backand-forth messages, in a noise-resilient way. Following Schulman's question regarding the maximum rate of interactive codes <ref type="bibr">[15]</ref>, Kol and Raz <ref type="bibr">[12]</ref> defined the notion of interactive channel capacity, which is the analogue of channel capacity in the interactive setting. For every &#120598; &gt; 0, they designed an interactive code with rate &#119903; &#120598; = 1 -O ( &#8730;&#65025; H(&#120598;)) over the two-way binary symmetric channel, under the assumption that the protocol being encoded is alternating <ref type="foot">2</ref> .</p><p>It is not hard to see that the interactive coding scheme of Kol and Raz <ref type="bibr">[12]</ref> also works for the C &#915;,&#120598; channel, for every &#915;. Their result, stated for such channels, is that for any &#120598; &gt; 0, any alphabet set &#915;, and any alternating protocol &#928; with alphabet &#915;, there exists a protocol &#928; &#8242; that simulates &#928; over C &#915;,&#120598; with negligible error, and has length |&#928;| &#8226; (1/&#119903; &#120598; + &#119900; (1)), where |&#928;| is the length of &#928;. Observe that, as in the classical setting, the maximum rate approaches 1 as &#120598; approaches 0. Following <ref type="bibr">[12]</ref>, the dependence of the maximum rate on &#120598;, under the same alternating turns assumption, was further improved by <ref type="bibr">[9]</ref> to 1 -O ( &#8730; &#120598;), and was also studied for other two-way channels, including the adversarial channel <ref type="bibr">[4,</ref><ref type="bibr">9]</ref>, the (adversarial) feedback channel <ref type="bibr">[7,</ref><ref type="bibr">14]</ref>, the adversarial erasure channel <ref type="bibr">[7]</ref>, and the adversarial insertion-deletion channel <ref type="bibr">[10]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Our Result</head><p>The main result of this paper is Theorem 1.1, showing that in the general case, where the order of speaking in the noiseless communication protocols &#928; may be arbitrary, the maximum achievable rate is bounded away from 1.</p><p>Theorem 1.1. For every &#120598; &gt; 0, there exists a set &#915; and a deterministic protocol &#928; with alphabet &#915;, such that any randomized protocol &#928; &#8242; that simulates &#928; over C &#915;,&#120598; with probability 0.99 has length at least |&#928;|/(1 -&#937;(1)).</p><p>Observe that since Theorem 1.1 holds for the C &#915;,&#120598; channel that has stochastic noise and for public-coin protocols &#928; &#8242; , it also holds for adversarial noise and private-coin protocols. Furthermore, our proof of Theorem 1.1 actually proves a much stronger claim (see <ref type="bibr">Section 2)</ref>. For example, it implies that the maximum rate of an interactive code over the feedback channel that randomly erases a single communicated symbol (i.e., one of the sent symbols, selected uniformly at random, is received as '&#8869;' and the sender is notified) is only 1 -&#937;(1) (cf. the results of <ref type="bibr">[4,</ref><ref type="bibr">7,</ref><ref type="bibr">9,</ref><ref type="bibr">10,</ref><ref type="bibr">12,</ref><ref type="bibr">14]</ref> for such channels with maximum rate approaching 1).</p><p>We mention that our result settles the (non-binary version) of a conjecture by Haeupler (Conjecture 1.1 in <ref type="bibr">[9]</ref>), that also appears in Haeupler and Gelles (Question 3 in Section 7 of <ref type="bibr">[7]</ref>) and in Gelles's excellent survey (Question 2 in Section 5 of <ref type="bibr">[6]</ref>). While lower bounds on the maximum rate of various two-way channels (i.e., upper bounds on the overhead of interactive codes) are known, prior to our work, the only non-trivial upper bound was due to <ref type="bibr">[12]</ref> and is extremely involved (see Section 1.2).</p><p>We also mention that Theorem 1.1 uses a large alphabet set (specifically, we need |&#915;| = poly(|&#928;|)), as for such alphabets the single erased symbol cannot be guessed by the receiver with high probability (in the binary |&#915;| = 2 case, the erased symbol can be guessed with probability 1  2 ). Nevertheless, we believe that Theorem 1.1 still holds for the binary setting (fixing &#915; = {0, 1}), and proving it is an outstanding question we leave open. Other interesting directions for future work include finding the maximum rate of interactive codes over C &#915;,&#120598; , say when &#120598; approaches 0, and characterizing the &#322;hard&#382; communication orders resulting in maximum rates bounded away from 1.</p><p>Finally, we wish to point out a corollary of Theorem 1.1: Many works involving interactive protocols (in the noisy or noiseless settings) assume an alternating order of speaking, as it is often simpler to deal with and only incurs at most a factor 2 blowup to the communication. Theorem 1.1 shows that this transformation of general protocols to alternating ones incurs at least a factor &#119888; blowup, for some &#119888; &gt; 1: Assume that the blowup is only by a 1 + &#119900; (1) factor. By converting the hard-to-simulate protocol &#928; from Theorem 1.1 to a protocol with alternating turns and then applying the <ref type="bibr">[12]</ref> scheme, we obtain a noise-resilient protocol &#928; &#8242; that simulates &#928; with only 1 + &#119900; (1) blowup to the communication, in contradiction to Theorem 1.1.</p><p>1.1.1 Techniques. The proof of Theorem 1.1 is quite involved and a detailed overview can be found in Section 2. In this section we give some of the highlights of our proof.</p><p>Theorem 1.1 is proved by combining Theorem 4.1 and Theorem 4.2. As mentioned above, our result holds even over the very mildly noisy channel that has feedback and only randomly erases a single communicated symbol. Theorem 4.1 considers a pointer chasing protocol with order of speaking &#120590;<ref type="foot">foot_2</ref> and shows that it can only be simulated over this mildly noisy channel by a protocol with order of speaking &#120590; &#8242; , for which &#120590; is a strong subsequence of &#120590; &#8242; . By a strong subsequence, we mean that for most coordinates &#119894; &#8242; of &#120590; &#8242; , &#120590; remains a subsequence of &#120590; &#8242; even after coordinate &#119894; &#8242; is removed. Observe that given Theorem 4.1, to prove Theorem 1.1, all we have to do is exhibit a &#120590; such that any &#120590; &#8242; for which &#120590; is a strong subsequence of &#120590; &#8242; is a constant factor longer than &#120590;. This is done in Theorem 4.2.</p><p>At a high level, Theorem 4.1 is proved by proving a generalized pointer chasing lower bound: while prior pointer chasing lower bounds assume that players alternate (e.g., <ref type="bibr">[13]</ref>), our proof holds for any order of speaking. To analyze cases where one of the parties speaks in several consecutive rounds, we use a lower bound for a generalization of the well-known Index problem, where the communication is not one-way, but the party holding the index speaks substantially less than what it takes to convey the index. To see why this lower bound is useful, assume for example that in the noiseless protocol Alice speaks three times and then Bob speaks once, i.e., &#120590; = &#119860;&#119860;&#119860;&#119861;. We think of Alice's message in those three rounds as an index &#119894;, and of Bob's input as a vector &#119907;. When Bob speaks in the fourth round he gives &#119907; &#119894; to Alice. Now consider a simulation protocol with order of speaking &#120590; &#8242; = &#119861;&#119860;&#119861;&#119860;&#119861;&#119860;. Can Bob give &#119907; &#119894; to Alice? We show that he cannot. The reason is that Alice only speaks in two instead of three rounds before Bob's final round, thus she can only give partial information about &#119894;, which is not enough for Bob to compute &#119907; &#119894; . Theorem 4.2 is a purely combinatorial claim about strong subsequences, and is shown using the probabilistic method. We provide a detailed overview in Section 2.2, but for the high level idea, consider, for any &#119879; &gt; 0 the pair of strings (&#120590;, &#120590; &#8242; ) = (&#119860;&#119861;) &#119879; , (&#119860;&#119861;) &#119879; +1 , and observe that &#120590; is a strong subsequence of &#120590; &#8242; and &#120590; &#8242; is almost the same length as &#120590;. Roughly speaking, our proof shows that the only reason &#120590; is a subsequence of &#120590; &#8242; for such a short &#120590; &#8242; , is that &#120590; is highly &#322;predictable&#382;, in the sense that one can &#322;guess&#382; the symbols after coordinate &#119894; based on the previous symbols. We formalize this notion and show that a uniformly random &#120590; is not predictable, and use this to show that for most &#120590;, no short &#120590; &#8242; will be such that &#120590; is a strong subsequence of &#120590; &#8242; .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">Additional Related Work</head><p>We next survey the most relevant work on the maximum rates of interactive codes over different channels.</p><p>The C {0,1},&#120598; channel. The study of error correcting codes for interactive communication was pioneered by Schulman <ref type="bibr">[15]</ref>, who showed how to transform any interactive communication protocol over the (noiseless) binary channel to an equivalent noise-resilient protocol that works over the (two-way) C {0,1},&#120598; channel, with only a constant overhead in the communication. This shows that for any &#120598; &lt; 1  2 , the maximum rate of an interactive code over C {0,1},&#120598; is at least some constant strictly greater than 0.</p><p>Kol and Raz <ref type="bibr">[12]</ref> studied the maximum rate &#119903; &#120598; achievable by any interactive code over C {0,1},&#120598; , but as mentioned above, it is not hard to see that their results hold for every channel C &#915;,&#120598; . They showed that for alternating noiseless protocols and protocols whose communication order is periodic with a small period, &#119903; &#120598; = 1-&#119874; &#8730;&#65025; H(&#120598;) . The assumption that the noiseless protocol is alternating (or has a small period) is crucial as their coding scheme uses the rewind-iferror mechanism <ref type="bibr">[15]</ref>, where the parties run the noiseless protocol over the noisy channel, and periodically compare their received transcripts to detect errors. If an error was detected, the parties &#322;rewind&#382; to the last agreed upon point and continue the execution of the noiseless protocol from that point. Since the noiseless protocol is assumed to be alternating, by taking the order of speaking of the simulating protocol to also be alternating, they can ensure that when rewinding, the order of speaking in the simulation matches the assumed order of speaking in the noiseless protocol. Kol and Raz also proved a matching upper bound of 1 -&#937; &#8730;&#65025; H(&#120598;) for some carefully chosen communication orders 4 . We mention that the Kol-Raz result gives the first separation between the maximum rate of classical error correcting codes and that of interactive codes, and observe that Theorem 1.1 gives a substantially stronger separation.</p><p>Building on <ref type="bibr">[12]</ref> and also assuming an alternating order of speaking, <ref type="bibr">[8]</ref> presented a deterministic coding scheme that achieves a rate of &#119903; &#120598; (the <ref type="bibr">[12]</ref> scheme is randomized), and <ref type="bibr">[2]</ref> gave a coding scheme that handles larger &#120598;'s (observe that the <ref type="bibr">[12]</ref> scheme is only meaningful for small &#120598;'s). Specifically, <ref type="bibr">[2]</ref> showed that the maximum rate of interactive codes over C {0,1},&#120598; is at least 0.0302 times the maximum rate of classical error correcting codes over C {0,1},&#120598; .</p><p>Other (non-adaptive) channels. The maximum rates of interactive codes over other two-way channels, that are well studied in the context of classical codes, were also considered with the alternating communication order assumption. Pankratov <ref type="bibr">[14]</ref> studied the rate of interactive codes over channels with random errors and feedback, and gave a scheme with rate 1 -&#119874; ( &#8730; &#120598;). Haeupler and Gelles <ref type="bibr">[7]</ref> improved his result and gave a scheme with rate 1 -&#920;(H(&#120598;)) that works for the adversarial feedback channel. A scheme with the same rate was also given by <ref type="bibr">[7]</ref> for the adversarial erasure channel. The adversarial channel with corruption errors (bit flips) was considered in <ref type="bibr">[3,</ref><ref type="bibr">4,</ref><ref type="bibr">9,</ref><ref type="bibr">16]</ref>, and an interactive code with rate &#119903; &#120598; for this model was presented in <ref type="bibr">[4]</ref>. The adversarial binary insertion-deletion channel was considered by <ref type="bibr">[10]</ref>, who demonstrated an interactive coding scheme with rate &#119903; &#120598; for it.</p><p>Adaptive channels. In this paper as well as in most of the prior works on interactive coding, including all the paper surveyed so far, the assumption is that the protocols &#928; and &#928; &#8242; have a nonadaptive (a.k.a, oblivious or static) communication order, meaning 4 Specifically, the upper and lower bounds match when the communication order is periodic with a period &#119896; that satisfies &#120598; = &#920; log &#119896; &#119896; 2 . Indeed, Hauepler and Velingker <ref type="bibr">[11]</ref> showed that if the parties alternate in sending &#119896; = &#937; (poly(1/&#120598; ) ) consecutive symbols, then the maximum rate is 1 -&#920;(H(&#120598; ) ), violating the upper bound of <ref type="bibr">[12]</ref>. that the order of communication in the protocol is fixed in advance. Haeupler <ref type="bibr">[9]</ref> considered the adaptive setting, where at any round, each party decides whether to send a bit or listen for one based on its input and received transcript (which, in turn, depends on the channel's noise). Observe, however, that protocols in these models may have several parties attempting to send a symbol in the same round, or even no senders at all, and the received bits in these cases need to be specified.</p><p>Haeupler <ref type="bibr">[9]</ref> constructed interactive codes that encode nonadaptive protocols &#928; (with any communication order) by adaptive protocols over the C {0,1},&#120598; channel with rate 1 -O &#8730; &#120598; , bypassing the upper bound of <ref type="bibr">[12]</ref>. Put together, <ref type="bibr">[12]</ref> and <ref type="bibr">[9]</ref> imply a separation between the maximum rates obtained via adaptive and non-adaptive encodings. <ref type="bibr">[9]</ref> conjectured that this separation can be strengthened, even for a single erasure error, and our Theorem 1.1 proves his conjecture. We mention that other adaptive models were studied in the literature, see e.g., <ref type="bibr">[1,</ref><ref type="bibr">5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">OVERVIEW</head><p>Our main result (Theorem 1.1) says that regardless of how small the noise parameter &#120598; is, the overhead required to simulate a noiseless channel over a noisy channel that corrupts each symbol with probability &#120598; independently, is a constant strictly larger than 1. As mentioned in Section 1.1, we will actually prove a much stronger version of this, showing that it holds even if the channel corrupts exactly one (uniformly chosen) round, and the parties know in advance which round it is (and therefore, can change the simulation protocol they use arbitrarily based on this round, as long as this change does not affect the order in which the parties speak in the other rounds).</p><p>Showing a lower bound for a simulation protocol in a noise model that allows the protocol to change in response to the noise essentially means that we have to show a lower bound for a noiseless protocol, where all we know about the noiseless protocol is that its order of speaking is the same regardless of which round is corrupted by noise. Thus, a big part of our proof (Section 5) is, given two orders of speaking &#120590;, &#120590; &#8242; , understanding when can noiseless protocols with order of speaking &#120590; &#8242; simulate noiseless protocols with order of speaking &#120590;. This part subsumes and generalizes famous &#322;pointerchasing&#382; and &#322;round-complexity&#382; lower bounds in communication complexity <ref type="bibr">[13, e.</ref>g.] and is overviewed in Section 2.1. The answer turns out to be quite elegant: &#120590; &#8242; can simulate &#120590; if and only if &#120590; is a subsequence of &#120590; &#8242; .</p><p>We now look back at our original problem of designing a noiseless protocol &#928; that cannot be simulated by any (short) protocol over a noisy channel, even when the noise corrupts only one random symbol in the simulation protocol that is known to the parties as soon as they fix the order of speaking in the simulation protocol. Having shown that an order &#120590; &#8242; can simulate &#120590; if and only if &#120590; is a subsequence of &#120590; &#8242; , this means that we have to construct an order of speaking &#120590; (which will be the order in which the parties speak in &#928;) such that any short order of speaking &#120590; &#8242; satisfies the property that &#120590; is not a &#322;strong&#382; subsequence of &#120590; &#8242; . By that we mean that removing one uniformly chosen coordinate from &#120590; &#8242; ensures that, with high probability, &#120590; is not a subsequence of &#120590; &#8242; with that coordinate removed (see <ref type="bibr">Definition 3.7)</ref>. This is the second main part of our proof and is described in Section 6 and overviewed in Section 2.2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Lower Bounds on Noiseless Simulations</head><p>Recall that the order of speaking for a protocol &#928; of length &#119879; is a string &#120590; &#8712; {&#119860;, &#119861;} &#119879; that captures the order in which Alice and Bob speak in &#928;, in the sense that, for all &#119894; &#8712; [&#119879; ], party &#120590; &#119894; is the party speaking in round &#119894; of &#928;. The goal of this section is to show that, given any two orders of speaking &#120590; and &#120590; &#8242; , all (noiseless) protocols with order of speaking &#120590; can be simulated by (noiseless) protocols with order of speaking &#120590; &#8242; if and only if &#120590; is a subsequence of &#120590; &#8242; . The &#322;if&#382; direction is straightforward and we shall focus on showing the &#322;only if&#382; direction.</p><p>We argue this in the contrapositive. Suppose that two orders &#120590; and &#120590; &#8242; are given such that &#120590; is not a subsequence of &#120590; &#8242; . We first note that if &#120590; is alternating, i.e., &#120590; is of the form &#119860;&#119861;&#119860;&#119861;&#119860; . . . , then the desired result follows from (an easy extension of) the pointer chasing lower bounds in <ref type="bibr">[13]</ref> and subsequent work. However, a lower bound only for alternating &#120590; is not good enough for us, as we want the lower bound for a string &#120590; such that any short &#120590; &#8242; satisfies the property that &#120590; is not a strong subsequence of &#120590; &#8242; . This is provably not the case for alternating &#120590; as for any alternating &#120590;, the string &#120590; &#8242; = &#119860;&#119861;||&#120590; satisfies the property that &#120590; is a strong subsequence of &#120590; &#8242; , where || denotes concatenation.</p><p>However, as our lower bound must subsume these lower bounds, it is important to understand them. For this, consider the case when &#120590; = &#119860;&#119861; so that &#120590; &#8242; (as &#120590; cannot be a subsequence of &#120590; &#8242; ) is of the form &#119861;&#119861; . . . &#119860;&#119860;&#119860;&#119860; . . . , say &#120590; &#8242; = &#119861;&#119861;&#119861;&#119860;&#119860;&#119860;. Consider now the well-known Index problem, where Bob has a large array and Alice has an index for the array, and the goal of the parties is to output the element at Alice's index in Bob's array. There is a simple protocol with order of speaking &#120590; that solves this problem, where Alice first sends her index and then Bob sends the element at that index. However, if the order of speaking is restricted to be &#120590; &#8242; there is no way for Bob to send the right element to Alice, as all his messages are before he acquires any knowledge of Alice's index (unless of course, he sends to Alice the entire array, but this is impossible if Alice's alphabet is large enough).</p><p>For our more general result, we first extend the above lower bound to a more general class of &#120590; that has many Alice messages before the last Bob message, say, &#120590; = &#119860;&#119860;&#119860;&#119861;. The hard protocol for these &#120590; is also the protocol for the Index problem except that this time, Alice's index is so large that it will not fit in one message (and requires three messages). For all &#120590; &#8242; where all Bob's messages precede all Alice's messages, the argument is the same as before, but this time there are additional &#120590; &#8242; that are not of the form above and satisfy that &#120590; is not a subsequence of &#120590; &#8242; , for example &#120590; &#8242; = &#119861;&#119860;&#119861;&#119860;&#119861;&#119860;.</p><p>When &#120590; &#8242; = &#119861;&#119860;&#119861;&#119860;&#119861;&#119860;, as a protocol with order of speaking &#120590; &#8242; proceeds, Bob does get some messages from Alice (in rounds 2 and 4) but these messages are not long enough to contain her entire index. Thus, to show a lower bound for such &#120590; &#8242; , we need to extend the aforementioned lower bound for Index to work for protocols where Bob has partial information about Alice's index. This is exactly what we do, showing that such partial information from Alice cannot help Bob in guessing the right index a whole lot, and he still cannot send her the right index without sending a huge portion of his array. However, Bob cannot send a huge portion without having high communication, which is impossible if &#120590; &#8242; is not much longer than &#120590;.</p><p>To extend this argument to general &#120590; and &#120590; &#8242; such that &#120590; is not a subsequence of &#120590; &#8242; , we break the string &#120590; into &#322;intervals&#382;, where an interval is defined a set of consecutive rounds where the same party is speaking, e.g., the first three Alice rounds in &#120590; = &#119860;&#119860;&#119860;&#119861;. For each such interval starting from the first, we treat it like the Index problem above, and show that the interval cannot be simulated unless the party speaking in that interval has spoken enough times in the simulation. Once the party has spoken enough times, we remove the interval from &#120590; and the corresponding rounds from &#120590; &#8242; and arrive at a smaller problem with a fewer number of intervals. As &#120590; is not a subsequence of &#120590; &#8242; , we will run out of rounds in &#120590; &#8242; before we run out of intervals in &#120590;, giving us a trivial protocol for a non-trivial task, a contradiction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Analysis of Strong Subsequences</head><p>In this part, our goal is to show that there exists an order of speaking &#120590; &#8712; {&#119860;, &#119861;} * , such that for any &#120590; &#8242; &#8712; {&#119860;, &#119861;} * for which &#120590; is a strong subsequence of &#120590; &#8242; , it holds that &#120590; &#8242; is a constant factor longer than &#120590;. Recall that &#120590; is a strong subsequence of &#120590; &#8242; if, for most coordinates &#119894; of &#120590; &#8242; , it holds that &#120590; is a subsequence of &#120590; &#8242; with coordinate &#119894; removed. Throughout this section, we will disregard the connection of &#120590; and &#120590; &#8242; to communication protocols, and look at them simply as strings in {&#119860;, &#119861;} * . Also, we let &#119879; be the length of &#120590; and assume that the length of &#120590; &#8242; is &#119879; &#8242; = (1 + &#120575;)&#119879; , where &#120575; &gt; 0 is a small constant.</p><p>Patterns. We will show this using the probabilistic method, categorizing the relevant pairs (&#120590;, &#120590; &#8242; ) into various &#322;patterns&#382;, where for each pattern &#120588; and each &#120590;, there is exactly one &#120590; &#8242; such that the pair (&#120590;, &#120590; &#8242; ) is in the pattern &#120588;. We then show that, for every fixed pattern, and a randomly chosen pair (&#120590;, &#120590; &#8242; ) in the pattern, the probability that &#120590; is a strong subsequence of &#120590; &#8242; is extremely small, small enough to union bound over all the patterns, and our result follows.</p><p>Specifically, a pattern for us will be defined by a string &#120588; &#8712; {&#119860;, &#119861;, &#8226;} &#119879; &#8242; such that the number of coordinates of &#120588; that are equal to the &#322;bullet&#382; symbol &#8226; is &#119879; . We say that a pair (&#120590;, &#120590; &#8242; ) is in the pattern &#120588; if it holds that upon &#322;inserting&#382; the string &#120590; in the bullet coordinates of &#120588;, we get the string &#120590; &#8242; . Note that we can indeed restrict attention to the pairs (&#120590;, &#120590; &#8242; ) that are in some pattern, as if a pair is not in any pattern, then it must be the case that &#120590; is not a subsequence of &#120590; &#8242; , and therefore, it cannot be a strong subsequence of &#120590; &#8242; either. Moreover, the number of patterns &#120588; is at most</p><p>&#8226;&#119879; and we will ensure that, assuming &#120575; is a small enough constant, the relevant probabilities are small enough for a union bound over all the patterns.</p><p>Analyzing a toy pattern. Following the above framework, we now fix a pattern &#120588; and show that for a random pair (&#120590;, &#120590; &#8242; ) in &#120588;, we have that &#120590; is not a strong subsequence of &#120590; &#8242; with high probability. The high level idea here is best understood by taking &#120588; = &#8226; &#119879; , the &#119879; -length string each of whose coordinates are &#8226;, even though this is not a valid pattern according to the definition above. However, picking &#120588; = &#8226; &#119879; also means that the only way a pair (&#120590;, &#120590; &#8242; ) can be in this pattern is if &#120590; = &#120590; &#8242; , implying that &#120590; is trivially not a strong subsequence of &#120590; &#8242; = &#120590;. Thus, to make the argument non-trivial, we will show that, for a uniformly random &#120590; &#8712; {&#119860;, &#119861;} &#119879; , even a prefix of &#120590; of length 0.9&#119879; is not a strong subsequence of &#120590;.</p><p>For this, consider what happens if we erase the first coordinate of &#120590; to get a string &#120590; -1 , and try to estimate the length of the longest prefix of &#120590; that is a subsequence of &#120590; -1 (the case when a different coordinate is erased is similar). To estimate the length of the longest prefix, we consider the greedy algorithm &#322;matching&#382; the string &#120590; to the string &#120590; -1 : Namely, match each coordinate of &#120590; to the earliest coordinate possible <ref type="foot">5</ref> in &#120590; -1 . To analyze this algorithm, for all &#119894; &#8712; [&#119879; ], define lag &#119894; to be the difference between &#119894; and the coordinate in &#120590; corresponding to the coordinate in &#120590; -1 that &#119894; is matched to. For example, if coordinate 1 of &#120590; is matched to the first coordinate in &#120590; -1 (equivalently, the second coordinate in &#120590;), then,</p><p>As &#120590; is uniformly random in {&#119860;, &#119861;} &#119879; , each coordinate of &#120590; is uniformly and independently random in {&#119860;, &#119861;}, and thus each coordinate in &#120590; will take (in expectation) two coordinates of &#120590; -1 to find a match. This means that, in expectation <ref type="foot">6</ref> , we have lag &#119894; &#8805; lag &#119894; -1 +1. Using concentration bounds, we can conclude that, except with probability exponentially small in &#119879; , at most a 0.9 fraction of the coordinates will end up being matched, implying that the length of the longest prefix of &#120590; that is a subsequence of the resulting string &#120590; -1 is at most 0.9&#119879; , as desired.</p><p>Towards actual patterns. The argument above does not extend to actual patterns &#120588; &#8712; {&#119860;, &#119861;, &#8226;} &#119879; &#8242; , but for a very specific reason: To understand the reason, note that the argument above crucially relied on the fact that lag is non-zero throughout (in fact, it starts from 1 and never decreases). This means that we are always trying to compare a coordinate in &#120590; to another &#322;fresh&#382; coordinate, which is independently and uniformly random, and this allowed us to say that lag increases by 1 in expectation. In fact, if the lag were to have been 0, then we would be comparing every coordinate in &#120590; to itself, which means that it will aways match, and we will therefore be able to match all of &#120590;. Now, observe that the presence of non-bullet coordinates in &#120588; can actually decrease the lag and make it 0, ruining our argument above. For an example, consider the case &#119879; &#8242; = &#119879; + 2, and the pattern &#120588; = &#8226;, &#119860;, &#119861;, &#8226; &#119879; -1 . Specifically, consider the case where the first coordinate is erased, creating a lag of 1. However, as the two coordinates &#119860;, &#119861; immediately follow the erased coordinate, one can always match the first coordinate of &#120590; to one of these coordinates, bringing the lag back down to 0, and allowing the rest of &#120590; to be matched as is.</p><p>To get around this, we use the observation that any non-bullet symbol in &#120588; can decrease the lag by at most 1. Thus, as the number of non-bullet symbols in &#120590; is &#120575;&#119879; , if we could somehow magically start with lag = &#120575;&#119879; , then, lag will never vanish for any fixed pattern &#120588; and we can apply exactly the same analysis as in the toy pattern above to get that &#120590; will not be a subsequence except with probability exponentially small in &#119879; . This probability is small enough for us to union bound over all possible &#120588; and get that except with probability exponentially small in &#119879; , a uniformly random &#120590; will not have any &#120590; &#8242; such that &#120590; is a strong subsequence of &#120590; &#8242; .</p><p>Starting with a small lag. All we need to do now is to apply the argument above for large initial lag to the case at hand where the initial lag is 1. For this, observe from our example above (and also in the toy example &#120588; = &#8226; &#119879; ) that lag will actually increase as &#119894; increases, in the sense that the final lag is at least &#937;(&#119879; ) more than the initial lag, except with exponentially small probability in &#119879; . This holds despite the fact that we have a small number (= &#120575;&#119879; ) of non-bullet symbols, as each such symbol can decrease lag by at most 1, but the much larger number (= &#119879; ) of bullet symbols, each increasing lag by 1 in expectation (as the small number of non-bullet symbols are not enough to make lag vanish), will eventually override the effect of the non-bullet symbols.</p><p>The fact that the lag increases can be used as follows: Suppose we are currently considering an &#119894; such that lag &#119894; is some value &#119871; &gt; 0. Consider the pattern starting from the coordinate where &#119894; is matched and look at the segment consisting of the next, say, &#119871;/ &#8730; &#120575; coordinates. As only a &#120575; fraction of the coordinates are nonbullet, this segment is expected to have &#8730; &#120575; &#8226; &#119871; &lt; &#119871; of non-bullet symbols. As the number of non-bullet symbols is smaller than the initial lag, we can conclude that (even after union bounding over all possible ways to place the non-bullet coordinates) except for a &#322;bad&#382; event that happens with probability exponentially small in &#119871;, this segment is expected to increase the lag to &#119871; &#8242; = &#937;(&#119871;/ &#8730; &#120575;). Now, we can consider the next segment of length &#119871; &#8242; / &#8730; &#120575;, and show that except with probability exponentially small in &#119871; &#8242; , this segment is expected to increase lag even more. The increasing length of these segments allows us to show that the sum of the bad probabilities converges to a constant, implying that for any erased coordinate &#119894;, one of the following holds: (1) The fraction of non-bullet symbols in one of the segments that are generated is much larger than &#120575;. (2) There exists a segment generated from &#119894; for which the bad event occurs. (3) When &#119894; is erased, the final lag increases to be &#937;(&#119879; ).</p><p>By Markov's inequality, both Items 1 and 2 will happen for at most a small constant fraction of &#119894;. Thus, there exists a &#120590; such that for most &#119894;, Item 3 will occur, implying that, for any &#120590; &#8242; such that &#120590; is a subsequence of &#120590; &#8242; , we have that &#120590; is not a subsequence of &#120590; &#8242; -&#119894; . It follows that there exists a &#120590; such that for no &#120590; &#8242; is it the case that &#120590; is a strong subsequence of &#120590; &#8242; , as desired. In terms of organization, the definition of &#119894; in Item 2 is formalized in Definition 6.6 and the proof that there is a small number of such &#119894; is in Lemmas 6.7 and 6.8. Analogous statements about Item 1 can be found in Lemma 6.10 while the proof of Item 3 can be found in Section 6.6 (specifically, Lemma 6.12).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">MODEL AND PRELIMINARIES 3.1 Notation</head><p>For &#119899; &gt; 0, we use [&#119899;] to denote the set {1, 2, . . . , &#119899;}. For &#119886;, &#119887; &gt; 0, we use [&#119886;, &#119887;] to denote the set {&#119886;, &#119886; + 1, . . . , &#119887;}. Additionally, we use (&#119886;, &#119887;] to denote the set {&#119886; + 1, . . . , &#119887;}. The notations [&#119886;, &#119887;) and (&#119886;, &#119887;) are defined analogously.</p><p>Let &#931; be an alphabet set and &#119899; &gt; 0 be an integer. For a string &#119904; &#8712; &#931; &#119899; and a set &#119868; &#8834; [&#119899;], we use &#119904; &#119868; to denote the |&#119868; |-length string obtained by taking only those coordinates of &#119904; that are in &#119868; , e.g., we have (&#119860;&#119861;&#119860;&#119860;&#119861;) {1,3,4} = &#119860;&#119860;&#119860;. For &#119894; &#8712; [&#119899;], we sometimes abbreviate &#119904; {&#119894; } to &#119904; &#119894; , &#119904; <ref type="bibr">[&#119894; ]</ref> to &#119904; &#8804;&#119894; , and &#119904; [&#119899;]\{&#119894; } to &#119904; -&#119894; . We also use the notations &#119904; &lt;&#119894; , &#119904; &gt;&#119894; and &#119904; &#8805;&#119894; that are defined analogously. Whenever we have &#119862; &#8712; {&#119860;, &#119861;}, we use &#119862; to denote the unique element of {&#119860;, &#119861;} not equal to &#119862;.</p><p>Throughout this paper, we use sans-serif letters to denote random variables.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Embedding Strings</head><p>Let &#120590;, &#120590; &#8242; &#8712; {&#119860;, &#119861;} * and &#119879; = |&#120590; | and &#119879; &#8242; = |&#120590; &#8242; |. For &#119894; &#8712; {0} &#8746; [&#119879; ], we define the function Emb(&#120590;, &#120590; &#8242; , &#119894;) inductively by Emb(&#120590;, &#120590; &#8242; , 0) = 0 and as follows when &#119894; &gt; 0:</p><p>Note that the min above is taken over a finite non-empty set, and is therefore well-defined. We also define the set:</p><p>(</p><p>We say that &#120590; is a subsequence of</p><p>we have:</p><p>We also use the following lemmas throughout our paper. The formal proofs of Lemmas 3.2 to 3.6 are omitted for space and can be found in the full version of the paper. </p><p>Moreover, we also have: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Our Noisy Channel</head><p>Let &#915; be a set with |&#915;| &#8805; 2. A (deterministic) protocol with the alphabet set &#915; is defined by a tuple:</p><p>where: (1) &#119879; &gt; 0 is a parameter denoting the length of the protocol, (2) &#120590; &#8712; {&#119860;, &#119861;} &#119879; is a string that determines which party speaks when (i.e., for all &#119894; &#8712; [&#119879; ], party &#120590; &#119894; is the unique party speaking in round &#119895;), (3) X &#119862; for &#119862; &#8712; {&#119860;, &#119861;} is the input set of party &#119862;, (4) Y is the set of possible outputs of the protocol, ( <ref type="formula">5</ref>) For all &#119894; &#8712; [&#119879; ], &#119872; &#119894; : X &#120590; &#119894; &#215; &#915; &#119894; -1 &#8594; &#915; is a function that computes the message sent in round &#119894; based on the input of the party &#120590; &#119894; speaking in round &#119894; and the transcript &#8712; &#915; &#119894; -1 received by party &#120590; &#119894; in the first &#119894; -1 rounds, (6) out &#119862; : &#915; &#119879; &#8594; Y for &#119862; &#8712; {&#119860;, &#119861;} are functions that each player uses to compute the output from the transcript of the protocol. We suppress items on the right hand side of Eq. ( <ref type="formula">3</ref>) when they are clear from context. We use the notation spkrs(&#928;) = &#120590; and |&#928;| = &#119879; . We define a randomized protocol &#928; to be a distribution over deterministic protocols &#928; that all have the same value of &#119879; , &#120590;, X &#119860; , X &#119861; , Y . We define spkrs(&#928;) and |&#928;| to be the common value of spkrs(&#928;) and |&#928;| respectively.</p><p>Execution of a protocol. Let &#928; be a protocol as above and &#120598; &#8805; 0. We now describe how &#928; is executed over the channel C &#915;,&#120598; that corrupts each sent symbol (independently) to a uniformly random symbol in &#915; with probability &#120598;. To describe this execution, we let &#9733; be a special symbol not in &#915; indicating &#322;no noise&#382; and &#119873; &#8712; (&#915; &#8746; {&#9733;}) &#119879; be a noise vector such that for all &#119894; &#8712; [&#119879; ], the symbol &#119873; &#119894; = &#9733; with probability 1 -&#120598; and a uniformly random symbol from &#915; with probability &#120598; (independently for all &#119894;), so that &#119873; captures the noise inserted by the aforementioned channel. We shall abuse notation and use C &#915;,&#120598; to denote both the channel and the above distribution over noise vectors.</p><p>The execution begins with both parties &#119862; &#8712; {&#119860;, &#119861;} having input &#119909; &#119862; &#8712; X &#119862; and proceeds in &#119879; rounds, maintaining the invariant that before round &#119894; &#8712; [&#119879; ], both parties &#119862; &#8712; {&#119860;, &#119861;} have a partial transcript &#928; &#119862; &lt;&#119894; &#8712; &#915; &#119894; -1 . In round &#119894;, party &#120590; &#119894; computes the symbol &#120574; &#119894; = &#119872; &#119894; &#119909; &#120590; &#119894; , &#928; &#120590; &#119894; &lt;&#119894; , appends it to its own partial transcript, and sends it over the channel to the other party &#120590; &#119894; .</p><p>The noise &#119873; then acts on the symbol as follows: If &#119873; &#119894; = &#9733;, then the symbol is sent uncorrupted and party &#120590; &#119894; receives the symbol &#120574; &#119894; . Otherwise, we have &#119873; &#119894; &#8712; &#915; and party &#120590; &#119894; receives the symbol &#119873; &#119894; . In either case, party &#120590; &#119894; appends the received symbol to its partial transcript and the execution proceeds to the next round.</p><p>After the &#119879; rounds are over, each party &#119862; &#8712; {&#119860;, &#119861;} outputs out &#119862; &#928; &#119862; &#8804;&#119879; &#8712; Y. Note that this execution is entirely determined by the triple &#119909; &#119860; , &#119909; &#119861; , &#119873; , which we shall often write as (&#119883;, &#119873; ) using &#119883; to denote the pair of inputs &#119909; &#119860; , &#119909; &#119861; . This fact allows us to write &#928; &#119862; &#119894; (&#119883;, &#119873; ), &#928; &#119862; &#8804;&#119894; (&#119883;, &#119873; ), etc. to denote the corresponding value in the execution of &#928; in the presence of noise &#119873; when the inputs are &#119883; . For &#119862; &#8712; {&#119860;, &#119861;}, we also define the notation res &#119862; &#928; (&#119883;, &#119873; ) to denote the output of party &#119862; in the above execution and res &#928; (&#119883;, &#119873; ) = res &#119860; &#928; (&#119883;, &#119873; ), res &#119861; &#928; (&#119883;, &#119873; ) . We omit &#119873; from the above notations when &#120598; = 0 and the execution is noiseless, as in this case, &#119873; is always the vector with all coordinates equal to &#9733;. Note that, in this case, the transcripts for Alice and Bob are the same and we can omit the superscript &#119862; in the notation.</p><p>Simulations and hole simulations. Let &#915; be an alphabet set as above. Let &#928; and &#928; &#8242; be two randomized protocols with alphabet &#915; and with the same input sets X &#119860; , X &#119861; for Alice and Bob. For &#119901; &#8712; [0, 1] and &#120598; &#8805; 0, we say the protocol &#928; &#8242; simulates the protocol &#928; over the channel C &#915;,&#120598; with probability &#119901; if for all &#119909; &#119860; &#8712; X &#119860; , &#119909; &#119861; &#8712; X &#119861; , it holds that</p><p>Throughout this text, the protocol &#928; being simulated will be deterministic and we shall omit it from the subscript above. As our main result in a lower bound, the fact that &#928; is deterministic only makes our result stronger. As &#915; is determined by &#928;, we shall sometimes omit writing &#322;over the channel C &#915;,&#120598; &#382; when &#120598; = 0.</p><p>For our proof of Theorem 1.1, we actually work with a different (and weaker <ref type="foot">7</ref> ) notion of simulation that we call &#322;hole simulation&#382; and is defined as follows: Let &#120590; &#8242; &#8712; {&#119860;, &#119861;} * and &#928; be a protocol as above. For &#119901; &#8712; [0, 1], we say that &#120590; &#8242; hole-simulates &#928; with probability &#119901; if there exists a set</p><p>10 such that for all &#119894; &#8242; &#8712; &#119868; &#8242; , there exists a randomized protocol &#928; &#8242; &#119894; &#8242; with alphabet &#915; and the same input sets as &#928; that simulates the protocol &#928; (over the channel C &#915;,0 ) with probability &#119901; and satisfies spkrs &#928; &#8242; &#119894; &#8242; = &#120590; &#8242; -&#119894; &#8242; .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Pointer Chasing</head><p>Let &#119898;,&#119879; &#8712; N. We inductively define the function PC </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">PROOF OF THEOREM 1.1</head><p>The goal is section is to prove Theorem 1.1 assuming two other theorems that we shall prove in the following sections. We start by stating these two theorems. Theorem 4.2. For all &#119879; &gt; 0, there exists &#120590; &#8712; {&#119860;, &#119861;} &#119879; such that for all &#120590; &#8242; &#8712; {&#119860;, &#119861;} * such that &#120590; is a strong subsequence of &#120590; &#8242; , we have</p><p>We are now ready to prove Theorem 1.1 (assuming Theorems 4.1 and 4.2).</p><p>Proof of Theorem 1.1. Fix &#120598; &gt; 0 and assume that &#120598; &lt; 0.001 without loss of generality. Define &#119879; = 1 &#120598; 2 and let &#120590; be as promised by Theorem 4.2. Define &#915; = (200&#119879; ) 200 and &#928; = PC &#120590; . Let &#928; &#8242; be a randomized protocol that simulates &#928; with over the channel C &#915;,&#120598; with probability 0.99 and &#120590; &#8242; = spkrs(&#928;). As the proof is trivial otherwise, assume that |&#120590; &#8242; | = |&#928; &#8242; | &#8804; 5&#119879; . We claim that &#120590; &#8242; hole simulates PC &#120590; with probability 0.5. This finishes the proof as it implies using Theorem 4.1 that &#120590; is a strong subsequence of &#120590; &#8242; which using Theorem 4.2 means that |&#120590; &#8242; | &#8805; 1 + 10 -100 &#8226; &#119879; , as desired.</p><p>It remains to show the claim. To this end, let &#119879; &#8242; = |&#120590; &#8242; | and for &#119894; &#8242; &#8712; [&#119879; &#8242; ], define the protocol &#928; &#8242; &#119894; &#8242; with spkrs(&#928; &#8242; &#119894; &#8242; ) = &#120590; &#8242; -&#119894; &#8242; as follows: For &#119905; &#8712; {0} &#8746; [&#119879; &#8242; ], let &#119901; &#119905; = &#119879; &#8242; &#119905; &#8226; &#120598; &#119905; (1 -&#120598;) &#119879; &#8242; -&#119905; be the probability that the channel C &#915;,&#120598; corrupts exactly &#119905; symbols in &#928; &#8242; . This means that &#119901; &#119905; 1-&#119901; 0 is the probability the channel C &#915;,&#120598; corrupts exactly &#119905; symbols in &#928; &#8242; conditioned on it corrupting at least one symbol. Then, the protocol &#928; &#8242; &#119894; &#8242; is exactly the same as &#928; &#8242; except that it (1) It does not have round &#119894; &#8242; and the party supposed to receive in this round assumes it got a uniformly random symbol in &#915;. Observe that this can equivalently be seen as the channel always corrupting round &#119894; &#8242; in &#928; &#8242; . (2) Samples &#119905; &#8242; &#8712; [&#119879; &#8242; ] with probability &#119901; &#119905; &#8242; 1-&#119901; 0 , and then artificially corrupts &#119905; &#8242; -1 rounds, ignoring the bit actually received in these rounds and using a uniformly random symbol in &#915; instead.</p><p>Observe that picking &#119894; &#8242; &#8712; [&#119879; &#8242; ] uniformly at random and running &#928; &#8242; &#119894; &#8242; over the noiseless channel C &#915;,0 is the same as running &#928; &#8242; over C &#915;,&#120598; and conditioning on the fact that the channel corrupts at least one symbol. As &#928; &#8242; simulates &#928; with over the channel C &#915;,&#120598; with probability 0.99, we get:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>It follows that Pr</head><p>10 values of &#119894; &#8242; , as desired. &#9633;</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">PROOF OF THEOREM 4.1</head><p>The goal of this section is to show Theorem 4.1. Owing to the definition of hole simulations and Definition 3.7, it suffices to show the following lemma: We prove Lemma 5.1 in the rest of this section. Fix &#120591;, &#928; and define &#119899; = |&#120591; |, &#119879; = |&#928;|, and &#120590; = spkrs(&#928;). We shall show the lemma in the contrapositive, assuming that &#120591; is a not subsequence of &#120590; and showing that &#928; does not simulate PC &#120591; with probability 1 &#119899; . As &#120591;, &#120590; are fixed we shall often omit them from our notation and write Emb(&#8226;) instead of Emb(&#120591;, &#120590;, &#8226;) and E instead of E(&#120591;, &#120590;).</p><p>Let X &#119860; and X &#119861; be input sets of Alice and Bob respectively in PC &#120591; (and therefore also in &#928;). Recall from Section 3.3 that we have to show that there exist &#119909; &#119860; &#8712; X &#119860; , &#119909; &#119861; &#8712; X &#119861; such that:</p><p>Let F be the uniform distribution over all inputs of PC &#120591; , defined as in Section 3.4. To show the foregoing equation, we fix an arbitrary deterministic protocol &#928; in the support of &#928; and show that (noting that res PC &#120591; (&#119865; ) = PC(&#119865; )):</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Notation</head><p>For a finite non-empty set &#119878;, we shall use U (&#119878;) to denote the uniform distribution over &#119878;. We omit &#119878; from the notation when it is clear from the context. All probabilities and random variables will be defined over the randomness in F , and we will often abbreviate Pr &#119865; &#8764;F to Pr for brevity of notation. Throughout, if X is a random variable and &#119909; is a value that X can take, we sometimes abbreviate the event X = &#119909; as simply &#119909; when it is clear from context. Thus, we may write Pr(&#119909;) instead of Pr(X = &#119909;) and Pr(&#8226; | &#119909;) instead of Pr(&#8226; | X = &#119909;). We use dist(X) to denote the distribution of a random variable X.</p><p>We will use F to denote the random variable corresponding to a sample from F and &#119865; to denote a given value of F. Observe that &#119865; is an &#119899;-tuple</p><p>We may combine these notations and use &#119891; &#119860; &#8804;&#119894; = &#119891; {&#119894; &#8242; &#8712; [&#119894; ] |&#120591; &#119894; &#8242; =&#119860;} , etc.. We will use f &#8804;&#119894; to denote the random variable corresponding to &#119891; &#8804;&#119894; . The notations f &#119878; , f &lt;&#119894; , f &#119860; , etc. are defined similarly.</p><p>Recall the functions &#928; &#119905; (&#8226;) and &#928; &#8804;&#119905; (&#8226;) from Section 3.3. In this section, we extend this notation to sets &#119878; &#8838; [&#119879; ] by defining &#928; &#119878; (&#8226;) = (&#928; &#119905; (&#8226;)) &#119905; &#8712;&#119878; . For &#119905; &#8712; [&#119879; ], we will use &#928; &#119905; = &#928; &#119905; (F) to denote the random variable obtained by sampling F and outputting &#928; &#119905; (F), and use &#928; &#119905; to denote a value &#928; &#119905; can take. The notations &#928; &#8804;&#119905; , &#928; &#119878; are defined analogously.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Definitions</head><p>Recall that we fixed &#120591; &#8712; {&#119860;, &#119861;} &#119899; as the order in which the parties speak in the protocol PC &#120591; being simulated. We also fixed a deterministic protocol &#928; and defined &#119879; = |&#928;| and &#120590; = spkrs(&#928;) &#8712; {&#119860;, &#119861;} &#119879; . The set L. We consider the set of indices &#119894; &#8712; [&#119899;] where the value of &#120591; &#119894; is different from &#120591; &#119894;+1 . Define the set:</p><p>Informally, L is the rounds where &#120591; &#322;switches&#382; from &#119860; to &#119861; or &#119861; to &#119860;. Equivalently, we partition &#120591; into consecutive intervals consisting of the same player and L is the set of endpoints of these intervals.</p><p>The element &#119899; is added to L for convenience. For &#119894; &#8712; [&#119899;], we define &#8467; &#8805; &#119894; to be smallest value &#8467; &#8712; L satisfying &#8467; &#8805; &#119894;. This is well defined as &#119899; &#8712; L is one such value. Similarly, define &#8467; &lt; &#119894; to be largest value &#8467; &#8712; {0} &#8746; L satisfying &#8467; &lt; &#119894;. Observe that for all &#119894; &#8712; [&#119899;], we have</p><p>Defining Good and Rem. For &#119905; &#8712; {0} &#8746; [&#119879; ], define the set</p><p>Informally, Good gets a round &#119905; of the protocol &#928; and outputs the first interval of &#120591; that we do not expect to have fully simulated after round &#119905;. Observe that Good(&#119905;)</p><p>and &#119894; &#8712; Good(&#119905;). Define:</p><p>Roughly speaking, Rem &#119894; (&#119905;) is the amount of (min-)entropy remaining in the random variable PC &gt;&#8467; &lt; &#119894; (f &#8804;&#119894; ) before round &#119905; of the protocol.</p><p>&#322;Revealing&#382; information. To make our analysis cleaner, we reveal some information to the players at various points in the protocol. More precisely, let &#119905; &#8242; &#8712; [3&#119879; ] be given and &#119865; in the support <ref type="foot">8</ref> of F be arbitrary. Let &#119905; &#8712; [&#119879; ] be the unique value satisfying 3(&#119905; -1) &lt; &#119905; &#8242; &#8804; 3&#119905;. We shall define values &#934; &#119905; &#8242; (&#119865; ) inductively. If &#119905; &#8242; = 3&#119905; -2, define:</p><p>If &#119905; &#8242; = 3&#119905; -1, define:</p><p>Informally, this definition amounts to revealing the correct transcript for any interval at the end of the interval. Finally, if &#119905; &#8242; = 3&#119905;, define:</p><p>Informally, this definition amounts to revealing, at the end of an interval, whether the right answer for the next interval can be guessed with probability much better than what Rem would indicate. We will later show that, this answer is no (= 0) with high probability. Henceforth, we treat &#934; the same way as &#928; in our notation, i.e., we let &#934; &#119905; &#8242; = &#934; &#119905; &#8242; (F) denote the random variable obtained by sampling F and outputting &#934; &#119905; &#8242; (F), use &#934; &#119905; &#8242; to denote a value &#934; &#119905; &#8242; can take, and define &#934; &#8804;&#119905; &#8242; , &#934; &#119878; etc. analogously to &#928; &#8804;&#119905; , &#928; &#119878; , etc. We claim that &#934; provides all the necessary information in order to reconstruct PC(f), as claimed below.</p><p>Claim 5.2. For any &#8467; &#8712; L satisfying Emb(&#8467;) &#8804; &#119879; , the value of &#934; &#8804;3Emb(&#8467; ) -1 fixes the value of PC(f &#8804;&#8467; ).</p><p>The formal proof of Claim 5.2 is omitted for space and can be found in the full version of the paper.</p><p>We also claim that &#934; is a transcript of a protocol. By this, we mean that each coordinate of &#934; can be computed fully by just one player using only their input and the transcript &#934; so far. Formally: Lemma 5.3. For &#119905; &#8242; &#8712; [3&#119879; ], there exists &#120590; &#8242; &#119905; &#8242; &#8712; {&#119860;, &#119861;} and a function &#119872; &#8242; &#119905; &#8242; such that for any &#119865; ,</p><p>Furthermore, for &#119905; &#8712; [&#119879; ], we have that:</p><p>The formal proof of Lemma 5.3 is omitted for space and can be found in the full version of the paper.</p><p>The sets Guess and Info. We are now ready to define the sets Guess and Info, the primary focus of our analysis. For &#119905; &#8712; {0} &#8746; [&#119879; ] and &#119894; &#8712; Good(&#119905;), we define:</p><p>Informally, this is the set of transcripts that allow us to guess the edges in the current interval (until &#119894;) with probability better than that indicated by Rem. For &#119905; &#8712; {0} &#8746; [&#119879; ] and &#119862; &#8712; {&#119860;, &#119861;}, define:</p><p>Informally, this is the set of transcripts that give a lot of information about party &#119862;'s input.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Properties of Info</head><p>This section is dedicated to proving Lemma 5.4, which will be key to proving our main result. Roughly, Lemma 5.4 says that transcripts are unlikely to be informative enough to be in the set Info as that requires information &#8805; &#119898; 0.01 which is much more than the communication (as &#119898; is larger than the communication to the power of 200).</p><p>Lemma 5.4. For all &#119905; &#8712; {0} &#8746; [&#119879; ] and &#119862; &#8712; {&#119860;, &#119861;},</p><p>For all &#119905; &#8242; &#8712; [3&#119879; ], let &#120590; &#8242; &#119905; &#8242; and &#119872; &#8242; &#119905; &#8242; be as in Lemma 5.3. We define for all &#119862; &#8712; {&#119860;, &#119861;}, for all &#119905; &#8242; &#8712; {0} &#8746; [3&#119879; ], for all &#934; &#8804;&#119905; &#8242; , the set:</p><p>we have</p><p>Roughly, our definition of &#934; ensures that the pairs of inputs that lead to the transcript &#934; &#8804;&#119905; &#8242; form a combinatorial rectangle (Rec denotes rectangle), and Rec &#119862; denotes the projection of this rectangle on party &#119862;'s inputs. In other words, Rec &#119862; is the set of all inputs of party &#119862; that may lead to the transcript &#934; &#8804;&#119905; &#8242; .</p><p>Observation 5.5. For all &#119905; &#8242; &#8712; [3&#119879; ], for all &#934; &#8804;&#119905; &#8242; , for all &#119862; &#8712; {&#119860;, &#119861;}, if &#120590; &#8242; &#119905; &#8242; &#8800; &#119862;, Rec &#119862; (&#934; &#8804;&#119905; &#8242; ) = Rec &#119862; (&#934; &lt;&#119905; &#8242; ). We now show several properties of Rec &#119862; . The formal proofs of Lemmas 5.6 to 5.8 are omitted for space and can be found in the full version of the paper.</p><p>Lemma 5.6. For all &#119905; &#8242; &#8712; {0} &#8746; [3&#119879; ], for all &#934; &#8804;&#119905; &#8242; , the event &#8704;&#119862; &#8712; {&#119860;, &#119861;} : f &#119862; &#8712; Rec &#119862; (&#934; &#8804;&#119905; &#8242; ) and the event &#934; &#8804;&#119905; &#8242; are equivalent.</p><p>Lemma 5.7. For all &#119905; &#8242; &#8712; {0} &#8746; [3&#119879; ], for all &#934; &#8804;&#119905; &#8242; ,</p><p>We now have the tools necessary to finish the proof of Lemma 5.4.</p><p>Proof of Lemma 5.4. For all &#934; &#8804;3&#119905; &#8712; Info &#119862; (&#119905;) we have:</p><p>(Eq. ( <ref type="formula">12</ref>))</p><p>The result follows from Lemma 5.8. &#9633;</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Key Lemma</head><p>We now show our key lemma.</p><p>Lemma 5.9. Let &#119905; &#8712; {0} &#8746; [&#119879; ] and &#119894; &#8712; Good(&#119905;). We have:</p><p>Proof. Proof by induction on &#119905;. The base &#119905; = 0 is straightforward as we get Emb &#8467; &lt; &#119894; = 0 which means that &#8467; &lt; &#119894; = 0 implying that Guess &#119894; (&#119905;) = &#8709;. We show the result for &#119905; &gt; 0 assuming it holds for smaller values of &#119905;. We consider the following cases: When Emb &#8467; &lt; &#119894; &lt; &#119905;: At a high level, this case amounts to analyzing a variant of the well-known Index problem, where the party holding the index can communicate a small number of bits but not enough to send the entire index. Let &#119906; = Emb &#8467; &lt; &#119894; for convenience. Note that &#119894; &#8712; Good(&#119906;). Applying the induction hypothesis on &#119906;, we get:</p><p>We assume &#120591; &#119894; = &#119860; as the argument for &#120591; &#119894; = &#119861; is analogous. For this, we shall fix an arbitrary &#934; &#8804;3&#119906; &#8713; Guess &#119894; (&#119906;), and an arbitrary &#119891; &#119861; and show that</p><p>By Eq. ( <ref type="formula">11</ref>), it suffices to show that:</p><p>We now focus on showing Eq. ( <ref type="formula">14</ref>). Define &#119911; = (0.2 + |E &#8745; (&#119906;, &#119905;]|) &#8226; log &#119898; for convenience. For this, we will apply Lemma A.7 with  <ref type="formula">9</ref>) and ( <ref type="formula">10</ref>) that for all &#119906; &#8242; &#8712; (&#119906;, &#119905;], we have &#934; 3&#119906; &#8242; -1 = &#934; 3&#119906; &#8242; = 0 and thus &#119892;(&#8226;) takes at most &#119898; |E&#8745;(&#119906;,&#119905; ] | many values. <ref type="foot">9</ref>From Lemma A.7, we get:</p><p>where:</p><p>Next, we claim that we can &#322;drop&#382; the conditioning on &#119891; &#119861; . For this, recall from Lemma 5.6 that the events &#934; &#8804;3&#119905; and &#8704;&#119862; &#8712; {&#119860;, &#119861;} : f &#119862; &#8712; Rec &#119862; (&#934; &#8804;3&#119905; ) are equivalent. As the latter event is a combinatorial rectangle (it is of the form (f &#119860; &#8712; A) &#8743; (f &#119861; &#8712; B) for some sets A, B) and the random variables f &#119860; and f &#119861; are independent, we get that the random variables f &#119860; and f &#119861; are also independent conditioned on &#934; &#8804;3&#119905; . Next, recall that PC &gt;&#8467; &lt; &#119894; (f &#8804;&#119894; ) is a function of f &#119860; conditioned on &#934; &#8804;3&#119906; , and conclude that PC &gt;&#8467; &lt; &#119894; (f &#8804;&#119894; ) and f &#119861; are also independent conditioned on &#934; &#8804;3&#119905; allowing us to drop &#119891; &#119861; . A similar argument allows us to drop &#119891; &#119861; from the other min-entropy term and we get:</p><p>By Eq. ( <ref type="formula">11</ref>) and our choice of &#934; &#8804;3&#119906; &#8713; Guess &#119894; (&#119906;), we have:</p><p>Using Eq. ( <ref type="formula">7</ref>) and the definition of &#119911;, we get:</p><p>This together with Eq. ( <ref type="formula">15</ref>) shows Eq. ( <ref type="formula">14</ref>).</p><p>When Emb &#8467; &lt; &#119894; = &#119905;: At a high level, the analysis in this case follows the popular pointer chasing lower bound of <ref type="bibr">[13]</ref>. We assume &#120591; &#119894; = &#119860; as the argument for &#120591; &#119894; = &#119861; is analogous. As &#119905; &gt; 0, we have &#8467; &lt; &#119894; &gt; 0 and we get &#120591; &#8467; &lt; &#119894; = &#119861;. It follows that &#120590; &#119905; = &#119861;. Let &#119906; = &#119905; -1. Applying the induction hypothesis on &#119906; and &#8467; &lt; &#119894; , we get:</p><p>By Lemma 5.4, we also have:</p><p>Thus, it suffices to show that:</p><p>For this, we shall fix an arbitrary &#934; &#8804;3&#119906; &#8713; Guess &#8467; &lt; &#119894; (&#119906;) &#8746; Info &#119860; (&#119906;) and show that:</p><p>For all &#119894; &#8242; &#8712; &#8467; &lt; &#119894; , &#8467; &#8805; &#119894; , define the set:</p><p>Roughly speaking, &#119878; is the set of prefixes &#119911; that allow the parties to guess the transcript in the next interval. Recall that Emb &#8467; &lt; &#119894; = &#119905; means that we are currently at the end of an interval. We now show that the probability of landing in &#119878; is small, as formalized in Eq. ( <ref type="formula">18</ref>) below. Next, use the fact that &#120591; &#119894; = &#119860; and the definition of &#8467; &lt; &#119894; and &#8467; &#8805; &#119894; to conclude that</p><p>. This, together with Lemma A.11 and the fact that</p><p>(Eq. ( <ref type="formula">17</ref>))</p><p>As such, we get that for all &#119894; &#8242; &#8712; &#8467; &lt; &#119894; , &#8467; &#8805; &#119894; , we have |&#119878; &#119894; &#8242; | &#8804; &#119898; 0.44 implying that |&#119878; | &#8804; &#119898; 0.45 . Next, note that as &#934; &#8804;3&#119906; &#8713; Guess &#8467; &lt; &#119894; (&#119906;), we also have by Eq. ( <ref type="formula">11</ref>) that H</p><p>Observe from Claim 5.2 that conditioning on &#934; &#8804;3&#119906; fixes the value of</p><p>It follows from Eq. ( <ref type="formula">7</ref>) and Definition A.5 that</p><p>As a consequence, Eq. ( <ref type="formula">16</ref>) follows if we show that:</p><p>Next, note from Claim 5.2 that the value of &#934; &lt;3&#119905; fixes the value of PC(f &#8804;&#8467; &lt; &#119894; ) (and also of &#934; &#8804;3&#119906; ). Thus, it suffices to fix an arbitrary &#934; &lt;3&#119905; that agrees with &#934; &#8804;3&#119906; and for which the corresponding value of PC(&#119891; &#8804;&#8467; &lt; &#119894; ) &#8713; &#119878;, and show that</p><p>By Eq. ( <ref type="formula">11</ref>), this is the same as:</p><p>Fixing such a &#934; &lt;3&#119905; , this is because of the following two claims. We omit the formal proofs of these claims, which can be found in the full version of the paper. Claim 5.10. It holds that:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.5">Finishing the Proof</head><p>We are now ready to prove Lemma 5.1</p><p>Proof of Lemma 5.1. Recall that we are showing the lemma in the contrapositive, assuming that &#120591; is a not subsequence of &#120590;. This means that Emb(&#119899;) &gt; &#119879; implying by Eq. ( <ref type="formula">6</ref>) that there exists &#119894; &#8712; [&#119899;] such that &#119894; &#8712; Good(&#119879; ). Fix such an &#119894; and apply Lemma 5.9 to conclude that Pr(E) &lt; 1 &#119899; 2 , where E is the event &#934; &#8804;3&#119879; &#8712; Guess &#119894; (&#119879; ). We now derive Eq. (4) as follows:</p><p>Thus, it suffices to show that Pr res &#928; (F) = PC(F) | E &#8804; 1 &#119899; 2 . We show this holds under a stronger conditioning by conditioning on an arbitrary &#934; &#8804;3&#119879; such that &#934; &#8804;3&#119879; &#8713; Guess &#119894; (&#119879; ). Fixing such a &#934; &#8804;3&#119879; and noting that fixing &#934; &#8804;3&#119879; also fixes res &#928; (F) to some value res we get:</p><p>(Eq. ( <ref type="formula">11</ref>) as &#934; &#8804;3&#119879; &#8713; Guess &#119894; (&#119879; ))</p><p>&#8804; &#119898; -0.5 (Eq. ( <ref type="formula">7</ref>))</p><p>In this section, we prove Theorem 4.2. For notational convenience, we define the constant &#120578; = 10 -5 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">A Customized Concentration Inequality</head><p>Fact 6.1. For all integers 1 &#8804; &#119896; &#8804; &#119899;, we have:</p><p>Lemma 6.2. Let &#119885; &#8838; N and &#119899; &gt; 0 be an integer. Also, let &#119901; &gt; 0 and X 1 , X 2 , &#8226; &#8226; &#8226; , X &#119899; be random variables taking values in N. Then, if &#120575; &gt; 0 is such that for all &#119894; &#8712; [&#119899;] and all &#119909; 1 , &#119909; 2 , &#8226; &#8226; &#8226; , &#119909; &#119894; -1 &#8712; N, we have:</p><p>Then, it holds that:</p><p>The formal proof of Lemma 6.2 is omitted for space and can be found in the full version of the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Basic Definitions</head><p>Recall that &#120578; = 10 -5 . Also recall from Section 2.2 that we consider segments of geometrically increasing lengths. These segments will be parameterized by an integer &#8467; &gt; 0. We will use &#119871; &#8467; to denote the length of segment &#8467;, &#119863; &#8467; to denote the &#322;delay&#382; or the &#322;lag&#382; before the segment starts, and &#119862; &#8467; to denote the non-bullet symbols in the pattern for this segment. We set these parameters as follows:</p><p>&#119862; &#8467; = &#120578; 6 &#119871; &#8467; &#119863; &#8467; = &#120578; 4 &#119871; &#8467; .</p><p>(19) We also define &#119871; &#8804;&#8467; = &#8467; &#8467; &#8242; =1 &#119871; &#8467; &#8242; and &#119871; &lt;&#8467; = &#8467; -1 &#8467; &#8242; =1 &#119871; &#8467; &#8242; . We adopt the convention that &#119871; &#8804;0 = 0 and observe that all these parameters integers. Next, we define the set {&#119860;, &#119861;} &#8226; to denote the set {&#119860;, &#119861;} &#8226; = {&#119860;, &#119861;} &#8746; {&#8226;}. For &#120588; &#8712; {&#119860;, &#119861;} * &#8226; , we use bull(&#120588;) to denote the number of coordinates in the string &#120588; that are equal to the &#322;bullet&#382; symbol 6.4 Strings With Small Pred &#8467; (&#8226;) Exist Lemma 6.8. For all integers &#119879; &gt; 0, there exists &#120590; &#8712; {&#119860;, &#119861;} &#119879; such that for all &#8467; &gt; 0, we have:</p><p>&#8226; &#119879; . The formal proof of Lemma 6.8 is omitted for space and can be found in the full version of the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.5">Structure of Long Subsequences</head><p>For the remainder of this section, readers may like to recall the definition of the set E &#120590;,&#120590; &#8242; in Eq. ( <ref type="formula">2</ref>). We borrow the following lemma from <ref type="bibr">[16]</ref>. Lemma 6.9 ( <ref type="bibr">[16]</ref>, <ref type="bibr">Lemma 6)</ref>. Let &#119879; &#8242; &gt; 0 be an integer. Also, let I be an indexing set and a collection of pairs &#119905; &#8242; &#119894; , &#119905; &#119894; &#119894; &#8712; I be given. Assume that 0 &#8804; &#119905; &#8242; &#119894; &lt; &#119905; &#119894; &#8804; &#119879; &#8242; for all &#119894; &#8712; I. There exists a set I &#8242; &#8838; I such that the intervals (&#119905; </p><p>The formal proof of Lemma 6.10 is omitted for space and can be found in the full version of the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.6">Proof of Theorem 4.2</head><p>Proof of Theorem 4.2. We define &#120590; to be the string promised by Lemma 6.8. Thus, for all &#8467; &gt; 0, we have:</p><p>Fix an arbitrary &#120590; &#8242; &#8712; {&#119860;, &#119861;} * such that &#120590; is a strong subsequence of &#120590; &#8242; and let &#119879; &#8242; = |&#120590; &#8242; |. Assume for the sake of contradiction that &#119879; &#8242; &lt; 1 + &#120578; 20 &#8226; &#119879; . As &#120590; is a strong subsequence of &#120590; &#8242; , we must have &#119879; &#8242; &#8805; &#119879; + 1 and E &#120590;,&#120590; &#8242; = &#119879; . From these, we conclude that 2&#119879; &#8805; &#119879; &#8242; &#8805; &#119879; &#8805; &#120578; -20 and E &#120590;,&#120590; &#8242; &#8805; 1 -&#120578;  and for all &#119894; &#8242; &#8712; &#119868; we have that &#120590; is a subsequence of &#120590; &#8242; -&#119894; &#8242; . Define the following sets: </p><p>We claim that: Claim 6.11. We have:</p><p>The formal proof of Claim 6.11 is omitted for space and can be found in the full version of the paper.</p><p>Conclude from Claim 6.11 and the fact that |&#119868; | &#8805; &#119879; &#8242; 10 that there exists an index &#119911; &#8242; &#8712; &#119868; \ &#119895; &gt;0 &#119868; &#119895; . We show that this leads to a contradiction. As &#119911; &#8242; &#8712; &#119868; , we have that &#120590; is a subsequence of &#120590; &#8242; -&#119911; &#8242; . Recall that this implies that E &#120590;, &#120590; &#8242; -&#119911; &#8242; = &#119879; or, equivalently, Emb &#120590;, &#120590; &#8242; -&#119911; &#8242; ,&#119879; &#8804; &#119879; &#8242; -1 &lt; &#119879; &#8242; . Henceforth, for notational convenience, we abbreviate Emb &#120590;, &#120590; &#8242; -&#119911; &#8242; , &#8226; to Emb * (&#8226;) and E &#120590;, &#120590; &#8242; -&#119911; &#8242; to E * . We also abbreviate Emb(&#120590;, &#120590; &#8242; , &#8226;) to Emb(&#8226;) and E(&#120590;, &#120590; &#8242; ) to E.</p><p>We now use the fact that &#119911; &#8242; &#8713; &#119895; &gt;0 &#119868; &#119895; to get more information about &#119911; &#8242; . From Eq. ( <ref type="formula">22</ref>), we get that &#119911; &#8242; &#8804; 0.999&#119879; &#8242; and &#119911; &#8242; &#8712; E. Due to Eq. ( <ref type="formula">2</ref>), this implies that there exists &#119911; &#8712; [&#119879; ] such that Emb(&#119911;) = &#119911; &#8242; . We claim that:</p><p>Indeed, if not, we have from Observation 3.1 that 0.999&#119879; &#8242; &#8805; &#119911; &#8242; &#8805; &#119911; &#8805; (1 -&#120578;) &#8226; &#119879; , a contradiction to &#119879; &#8242; &lt; 1 + &#120578; 20 &#8226; &#119879; . Next, Eq. ( <ref type="formula">22</ref>) also says that for all 0 &lt; &#119896; &#8804; &#119879; &#8242; -&#119911; &#8242; , we have (as the left hand side is an integer): To derive a contradiction, we claim that: Lemma 6.12. For all &#8467; &#8805; 0 such that &#119911; &#8804; &#119879; -3&#119871; &#8804;&#8467; , we have:</p><p>The formal proof of Lemma 6.12 is omitted for space and can be found in the full version of the paper.</p><p>We finish the proof of Theorem 4.2 by showing that it implies a contradiction. For this, define &#8467; * = log 10 10 &#120578; 4 &#119879; &#8242; and note that &#119879; &#8242; &#8805; &#120578; -20 implies that &#8467; * &#8805; 5. We get from Eq. ( <ref type="formula">19</ref>) that As we know that Emb * (&#119879; ) &lt; &#119879; &#8242; , this contradicts &#119879; &#8242; &lt; 1 + &#120578; 20 &#8226; &#119879; . &#9633;</p></div><note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0"><p>That is, the input and output alphabets of the channel C &#915;,&#120598; are &#915;, |&#915; | &#8805;</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1"><p>On a sent symbol &#119911; &#8712; &#915;, the channel outputs &#119911; with probability 1 -&#120598;, and with probability &#120598;, it outputs a random symbol in &#915;.<ref type="bibr">2</ref> That is, Alice sends a message to Bob in all odd rounds, and vice versa.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2"><p>We think of the order of speaking in a communication protocol as a string &#120590; &#8712; {&#119860;, &#119861; } * , where &#120590; &#119894; = &#119860; means that Alice speaks in round &#119894;, and &#120590; &#119894; = &#119861; means that Bob speaks in round &#119894;.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3"><p>For example, to match &#119860;&#119860;&#119861;&#119860; in the string &#119860;&#119861;&#119860;&#119861;&#119860;&#119861;, the matching will look like the following (matched characters underlined) &#119860;&#119861;&#119860;&#119861;&#119860;&#119861;.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4"><p>Note that if a match is found after &#119897; coordinates, the lag increases by &#119897; -1.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5"><p>The fact that we work with a weaker notion makes are proof stronger, and in particular, would also work for the erasure channel. To see the formal sense in which this is weaker, see Section 4.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6"><p>Henceforth, we omit writing &#322;in the support of&#382; when the random variable is clear from context.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_7"><p>For &#119905; &#8242; &#8712; (&#119906;, &#119905; ] where &#120591; &#119905; &#8242; = &#119861;, conditioning on &#119891; &#119861; makes each message &#934; 3&#119905; &#8242; -2 a deterministic function of the transcript so far. As such, there are at most &#119898; | {&#119905; &#8242; &#8712; (&#119906;,&#119905; ]:&#120591; &#119905; &#8242; =&#119860;} | possible transcripts, as there are &#119898; possible values of &#934; 3&#119905; &#8242; -2 for each &#119905; &#8242; &#8712; (&#119906;, &#119905; ] where &#120591; &#119905; &#8242; = &#119860;. Finally, by the definition of &#8467; &#8805; &#119894; and Eqs. (1) and (2), we get that {&#119905; &#8712; (&#119906;, &#119905; ] : &#120591; &#119905; &#8242; = &#119860;} = E &#8745; (&#119906;, &#119905; ].</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_8"><p>We adopt the convention that max( &#8709; ) = 0.</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_9"><p>Note that Eq. (19) implies that (1 -&#120578; ) &#8226; &#119871; &#8467; is an integer.</p></note>
		</body>
		</text>
</TEI>
