Changeset - 39a2a174b465
[Not reviewed]
0 1 0
András Gilyén - 8 years ago 2017-06-14 10:49:09
gilyenandras@gmail.com
Update on Overleaf.
1 file changed with 1 insertions and 1 deletions:
0 comments (0 inline, 0 general)
main.tex
Show inline comments
 
\documentclass[a4paper,11pt,english,final]{article}
 
\pdfoutput=1
 

	
 
\usepackage[utf8]{inputenc}
 
\usepackage[english]{babel}
 
\usepackage{fullpage}
 

	
 
\usepackage{graphics}
 
\usepackage{diagbox}
 
\usepackage[table]{xcolor}% http://ctan.org/pkg/xcolor
 
\usepackage{graphicx}
 
\usepackage{caption}
 
\captionsetup{compatibility=false}
 
\graphicspath{{./}}
 

	
 

	
 
\usepackage{tikz}
 
\usepackage{amssymb}
 
\usepackage{mathtools}
 
\usepackage{bm}
 
\usepackage{bbm}
 
%\usepackage{bbold}
 
\usepackage{verbatim}
 

	
 
%for correcting large brackets spacing
 
\usepackage{mleftright}\mleftright
 

	
 
\usepackage{algorithm}
 
\usepackage{algorithmic}
 
\usepackage{enumitem}
 
\usepackage{float}
 

	
 
%\usepackage{titling}
 

	
 
%\setlength{\droptitle}{-5mm}  
 

	
 
%\usepackage{MnSymbol}
 
\newcommand{\cupdot}{\overset{.}{\cup}}
 
\newcommand{\pvp}{\vec{p}{\kern 0.45mm}'}
 

	
 
\DeclarePairedDelimiter\bra{\langle}{\rvert}
 
\DeclarePairedDelimiter\ket{\lvert}{\rangle}
 
\DeclarePairedDelimiterX\braket[2]{\langle}{\rangle}{#1 \delimsize\vert #2}
 
\newcommand{\underflow}[2]{\underset{\kern-60mm \overbrace{#1} \kern-60mm}{#2}}
 

	
 
\def\Ind(#1){{{\tt Ind}({#1})}}
 
\def\Id{\mathrm{Id}}
 
\def\Pr{\mathrm{Pr}}
 
\def\Tr{\mathrm{Tr}}
 
\def\im{\mathrm{im}}
 
\newcommand{\bOt}[1]{\widetilde{\mathcal O}\left(#1\right)}
 

	
 
\newcommand{\QMAo}{\textsf{QMA$_1$}}
 
\newcommand{\BQP}{\textsf{BQP}}
 
\newcommand{\NP}{\textsf{NP}}
 
\newcommand{\SharpP}{\textsf{\# P}}
 

	
 
\newcommand{\diam}[1]{\mathcal{D}\left(#1\right)}
 
\newcommand{\paths}[1]{\mathcal{P}\left(#1\to\mathbf{1}\right)}
 
\newcommand{\maxgap}[1]{\mathrm{maxgap}\left(#1\right)}
 
\newcommand{\gaps}[1]{#1_{\mathrm{gaps}}}
 

	
 
\long\def\ignore#1{}
 

	
 
\newtheorem{theorem}{Theorem}
 
\newtheorem{corollary}[theorem]{Corollary}%[theorem]
 
\newtheorem{lemma}[theorem]{Lemma}
 
\newtheorem{prop}[theorem]{Proposition}
 
\newtheorem{definition}[theorem]{Definition}
 
\newtheorem{claim}[theorem]{Claim}
 
\newtheorem{remark}[theorem]{Remark}
 

	
 
\newenvironment{proof}
 
{\noindent {\bf Proof. }}
 
{{\hfill $\Box$}\\	\smallskip}
 

	
 
\usepackage[final]{hyperref}
 
\hypersetup{
 
	colorlinks = true,
 
	allcolors = {blue},
 
}
 
\usepackage{ifpdf} 
 
\ifpdf
 
	\typeout{^^J *** PDF mode *** } 
 
%	\input{myBiblatex.tex}
 
%	\addbibresource{LLL.bib}	
 
%\else
 
%	\typeout{^^J *** DVI mode ***} 
 
%	\hypersetup{breaklinks = true}
 
%	\usepackage[quadpoints=false]{hypdvips}
 
	\let\oldthebibliography=\thebibliography
 
	\let\endoldthebibliography=\endthebibliography
 
	\renewenvironment{thebibliography}[1]{%
 
		\begin{oldthebibliography}{#1}%
 
			\setlength{\itemsep}{-.3ex}%
 
	}%
 
	{%
 
		\end{oldthebibliography}%
 
	}
 
\fi 
 

	
 
%opening
 
\title{Criticality of resampling on the cycle / in the evolution model}
 
%\author{?\thanks{QuSoft, CWI and University of Amsterdam, the Netherlands. \texttt{?@cwi.nl} }
 
	%\and
 
	%?%
 
%}
 
%\thanksmarkseries{arabic}
 
%\renewcommand{\thefootnote}{\fnsymbol{footnote}}
 
%\date{\vspace{-12mm}}
 

	
 
\begin{document}
 
	
 
	\maketitle
 

	
 
	\begin{abstract}
 
		The model we consider is the following~\cite{ResampleLimit}: We have a cycle of length $n\geq 3$. Initially we set each site to $0$ or $1$ independently at each site, such that we set it $0$ with probability $p$. After that in each step we select a random vertex with $0$ value and resample it together with its two neighbours assigning $0$ with probability $p$ to each vertex just as initially. The question we try to answer is what is the expected number of resamplings performed before reaching the all $1$ state. 
 
		
 
		We present strong evidence for a remarkable critical behaviour. We conjecture that there exists some $p_c\approx0.62$, such that for all $p\in[0,p_c)$ the expected number of resamplings is bounded by a $p$ dependent constant times $n$, whereas for all $p\in(p_c,1]$ the expected number of resamplings is exponentially growing in $n$.
 
	\end{abstract}
 
	%Let $R(n)$ denote this quantity for a length $n\geq 3$ cycle.
 
	
 
	We can think about the resampling procedure as a Markov chain. To describe the corresponding matrix we introduce some notation. For $b\in\{0,1\}^n$ let $r(b,i,(x_{-1},x_0,x_1))$ denote the bit string which differs form $b$ by replacing the bits at index $i-1$,$i$ and $i+1$ with the values in $x$, interpreting the indices $\!\!\!\!\mod n$. Also for $x\in\{0,1\}^k$ let $p(x)=p((x_1,\ldots,x_k))=\prod_{i=1}^{k}p^{(1-x_i)}(1-p)^{x_i}$. Now we can describe the matrix of the Markov chain. We use row vectors for the elements of the probability distribution indexed by bitstrings of length $n$. Let $M_{(n)}$ denote the matrix of the leaking Markov chain:
 
	$$
 
		M_{(n)}=\sum_{b\in\{0,1\}^n\setminus{\{1\}^n}}\sum_{i\in[n]:b_i=0}\sum_{x\in\{0,1\}^3}E_{(b,r(b,i,x))}\frac{p(x)}{n-|b|},
 
	$$
 
	where $E_{(i,j)}$ denotes the matrix that is all $0$ except $1$ at the $(i,j)$th entry.
 

	
 
	We want to calculate the average number of resamplings $R^{(n)}$, which we define as the expected number of resamplings divided by $n$. For this let $\rho,\mathbbm{1}\in[0,1]^{2^n}$ be indexed with elements of $\{0,1\}^n$ such that $\rho_b=p(b)$ and $\mathbbm{1}_b=1$. Then we use that the expected number of resamplings is just the hitting time of the Markov chain:
 
	\begin{align*}
 
		R^{(n)}:&=\mathbb{E}(\#\{\text{resampling before termination}\})/n\\
 
		&=\sum_{k=1}^{\infty}P(\text{at least } k \text{ resamplings are performed})/n\\
 
		&=\sum_{k=1}^{\infty}\rho M_{(n)}^k \mathbbm{1}/n\\
 
		&=\sum_{k=0}^{\infty}a^{(n)}_k p^k
 
	\end{align*}
 

	
 
	\begin{table}[]
 
	\centering
 
	\caption{Table of the coefficients $a^{(n)}_k$}
 
	\label{tab:coeffs}
 
	\resizebox{\columnwidth}{!}{%
 
		\begin{tabular}{c|ccccccccccccccccccccc}
 
			\backslashbox[10mm]{$n$}{$k$} & 0 & 1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10 & 11 & 12 & 13 & 14 & 15 & 16 & 17 & 18 & 19 & 20 \\		\hline
 
			3 &	0 & 1 & \cellcolor{blue!25}2 & 3+1/3 & 5.00 & 7.00 & 9.33 & 12.00 & 15.00 & 18.33 & 22.00 & 26.00 & 30.33 & 35.00 & 40.00 & 45.333 & 51.000 & 57.000 & 63.333 & 70.000 & 77.000 \\
 
			4 &	0 & 1 & 2 & \cellcolor{blue!25}3+2/3 & 6.16 & 9.66 & 14.3 & 20.33 & 27.83 & 37.00 & 48.00 & 61.00 & 76.16 & 93.66 & 113.6 & 136.33 & 161.83 & 190.33 & 222.00 & 257.00 & 295.50 \\
 
			5 &	0 & 1 & 2 & 3+2/3 & \cellcolor{blue!25}6.44 & 10.8 & 17.3 & 26.65 & 39.43 & 56.48 & 78.65 & 106.9 & 142.2 & 185.8 & 238.7 & 302.41 & 378.05 & 467.13 & 571.14 & 691.69 & 830.44 \\
 
			6 &	0 & 1 & 2 & 3+2/3 & 6.44 & \cellcolor{blue!25}11.0 & 18.5 & 30.02 & 47.10 & 71.68 & 106.0 & 152.9 & 215.4 & 297.4 & 403.1 & 537.21 & 705.25 & 913.31 & 1168.2 & 1477.4 & 1849.1 \\
 
			7 &	0 & 1 & 2 & 3+2/3 & 6.44 & 11.0 & \cellcolor{blue!25}18.7 & 31.21 & 50.83 & 80.80 & 125.3 & 189.7 & 280.8 & 407.0 & 578.6 & 808.13 & 1110.2 & 1502.6 & 2005.6 & 2643.2 & 3443.1 \\
 
			8 &	0 & 1 & 2 & 3+2/3 & 6.44 & 11.0 & 18.7 & \cellcolor{blue!25}31.44 & 52.08 & 84.95 & 136.0 & 213.6 & 328.9 & 496.5 & 735.6 & 1070.7 & 1532.5 & 2159.5 & 2998.8 & 4108.1 & 5556.7 \\
 
			9 &	0 & 1 & 2 & 3+2/3 & 6.44 & 11.0 & 18.7 & 31.44 & \cellcolor{blue!25}52.30 & 86.27 & 140.7 & 226.3 & 358.4 & 558.4 & 855.4 & 1289.0 & 1911.5 & 2791.4 & 4017.2 & 5701.4 & 7985.9 \\
 
			10&	0 & 1 & 2 & 3+2/3 & 6.44 & 11.0 & 18.7 & 31.44 & 52.30 & \cellcolor{blue!25}86.49 & 142.1 & 231.6 & 373.4 & 594.8 & 934.4 & 1447.1 & 2209.0 & 3324.6 & 4934.8 & 7226.9 & 10447. \\
 
            \vdots \\
 
            15& 0 & 1 & 2 & 3+2/3 & 6.44 & 11.08 & 18.76 & 31.45 & 52.31 & 86.49 & 142.33 & 233.31 & 381.17 & 621.02 & 1009.38 & \cellcolor{blue!25}1637.13 & % 2650.74 & 4285.68 & 6913.55 & 11171.2 & 18052.2
 
            15& 0 & 1 & 2 & 3+2/3 & 6.44 & 11.08 & 18.76 & 31.45 & 52.31 & 86.49 & 142.33 & 233.31 & 381.17 & 621.02 & \cellcolor{blue!25}1009.38 & 1637.13 & % 2650.74 & 4285.68 & 6913.55 & 11171.2 & 18052.2
 
        \end{tabular}
 
	}
 
	\end{table}
 

	
 
	We observe that this is a power series in $p$. We discovered a very regular structure in this power series. It seems that for all $k\in\mathbb{N}$ and for all $n>k$ we have that $a^{(n)}_k$ is constant, this conjecture we verified using a computer up to $n=14$. 
 
	\newpage
 
	\noindent Based on our calculations presented in Table~\ref{tab:coeffs} and Figure~\ref{fig:coeffs_conv_radius} we make the following conjectures:
 
	\begin{enumerate}[label=(\roman*)]
 
		\item $\forall k\in\mathbb{N}, \forall n\geq 3 : a^{(n)}_k\geq 0$	\label{it:pos}	
 
        (A simpler version: $\forall k>0: a_k^{(3)}=(k+1)(k+2)/6$)
 
		\item $\forall k\in\mathbb{N}, \forall n>m\geq 3 : a^{(n)}_k\geq a^{(m)}_k$ \label{it:geq}		
 
		\item $\forall k\in\mathbb{N}, \forall n,m\geq \max(k,3) : a^{(n)}_k=a^{(m)}_k$ \label{it:const}		
 
  		\item $\exists p_c=\lim\limits_{k\rightarrow\infty}1\left/\sqrt[k]{a_{k}^{(k+1)}}\right.$ \label{it:lim}			
 
	\end{enumerate}
 
	We also conjecture that $p_c\approx0.61$, see Figure~\ref{fig:coeffs_conv_radius}.
 

	
 
	\begin{figure}[!htb]\centering
 
	\includegraphics[width=0.5\textwidth]{coeffs_conv_radius.pdf}
 
	%\includegraphics[width=0.5\textwidth]{log_coeffs.pdf}	
 
	\caption{$1\left/\sqrt[k]{a_{k}^{(k+1)}}\right.$} %$\frac{1}{\sqrt[k]{a_k^{(k+1)}}}$
 
	\label{fig:coeffs_conv_radius}
 
	\end{figure}
 
    
 
    For reference, we also explicitly give formulas for $R^{(n)}(p)$ for small $n$. We also give them in terms of $q=1-p$ because they sometimes look nicer that way.
 
    \begin{align*}
 
    	R^{(3)}(p) &= \frac{1-(1-p)^3}{3(1-p)^3}
 
        			= \frac{1-q^3}{3q^3}\\
 
    	R^{(4)}(p) &= \frac{p(6-12p+10p^2-3p^3)}{6(1-p)^4}
 
                    = \frac{(1-q)(1+q+q^2+3q^3)}{6q^4}\\
 
        R^{(5)}(p) &= \frac{p(90-300p+435p^2-325p^3+136p^4-36p^5+6p^6)}{15(1-p)^5(6-2p+p^2)}\\
 
                   &= \frac{(1-q)(6+5q+6q^2+21q^3+46q^4+6q^6)}{15q^5(5+q^2)}
 
    \end{align*}
 
    For $n=3$ the system becomes very simple because regardless of the current state, the probability of going to $111$ is always equal to $(1-p)^3$. Therefore the expected number of resamplings is simply the expectation of a geometric distribution. This gives the formula for $R^{(3)}(p)$ as shown above. Note that the $k$-th coefficient of the powerseries of a function $f(p)$ is given by $\frac{1}{k!}\left.\frac{d^k f}{dp^k}\right|_{p=0}$, i.e. the $k$-th derivative to $p$ evaluated at $0$ divided by $k!$. For the function $R^{(3)}(p) = (1-p)^{-3} - 1$ this yields $a^{(3)}_k = (k+2)(k+1)/6$ for $k\geq 1$ and $a^{(3)}_0=0$.
 

	
 
    ~
 

	
 
	If statements \ref{it:pos}-\ref{it:lim} are true, then we can define the function 
 
	$$R^{(\infty)}(p):=\sum_{k=0}^{\infty}a^{(k+1)}_k p^k,$$
 
	which would then have radius of convergence $p_c$, also it would satisfy for all $p\in[0,p_c)$ that $R^{(n)}(p)\leq R^{(\infty)}(p)$ and $\lim\limits_{n\rightarrow\infty}R^{(n)}(p)=R^{(\infty)}(p)$.
 
	It would also imply, that for all $p\in(p_c,1]$ we get $R^{(n)}(p)=\Omega\left(\left(\frac{p}{p_c}\right)^{n/2}\right)$.
 
	This would then imply a very strong critical behaviour. It would mean that for all $p\in[0,p_c)$ the expected number of resamplings is bounded by a constant $R^{(\infty)}(p)$ times $n$, whereas for all $p\in(p_c,1]$ the expected number of resamplings is exponentially growing in $n$.
 
	
 
	Now we turn to the possible proof techniques for justifying the conjectures \ref{it:pos}-\ref{it:lim}.
 
	First note that $\forall n\geq 3$ we have $a^{(n)}_0=0$, since for $p=0$ the expected number of resamplings is $0$.
 
	Also note that the expected number of initial $0$s is $p\cdot n$. If $p\ll1/n$, then with high probability there is a single $0$ initially and the first resampling will fix it, so the linear term in the expected number of resamplings is $np$, therefore $\forall n\geq 3$, $a^{(n)}_1=1$. 
 
	
 
	For the second order coefficients it is a bit harder to argue, but one can use the structure of $M_{(n)}$ to come up with a combinatorial proof. To see this, first assume we have a vector $e_b$ having a single non-zero, unit element indexed with bitstring $b$.
 
	Observe that $e_bM_{(n)}$ is a vector containing polynomial entries, such that the only indices $b'$ which have a non-zero constant term must have $|b'|\geq|b|+1$, since if a resampling produces a $0$ entry it also introduces a $p$ factor. Using this observation one can see that the second order term can be red off from $\rho M_{(n)}\mathbbm{1}+\rho M_{(n)}^2\mathbbm{1}$,
 
	which happens to be $2n$. (Note that it is already a bit surprising, form the steps of the combinatorial proof one would expect $n^2$ terms appearing, but they just happen to cancel each other.) Using similar logic one should be able to prove the claim for $k=3$, but for larger $k$s it seems to quickly get more involved.
 
	
 
	The question is how could we prove the statements \ref{it:pos}-\ref{it:lim} for a general $k$?
 
	
 
    \appendix
 
    
 
    \section{Lower bound on $R^{(n)}(p)$}
 
    Proof that \ref{it:pos} and \ref{it:lim} imply that for any fixed $p>p_c$ we have $R^{(n)}(p)\in\Omega\left(\left(\frac{p}{p_c}\right)^{n/2}\right)$. 
 
    
 
    By definition of $p_c = \lim_{k\to\infty} 1\left/ \sqrt[k]{a_k^{(k+1)}} \right.$ we know that for any $\epsilon$ there exists a $k_\epsilon$ such that for all $k\geq k_\epsilon$ we have $a_k^{(k+1)}\geq (p_c + \epsilon)^{-k}$. Now note that $R^{(n)}(p) \geq a_{n-1}^{(n)}p^{n-1}$ since all terms of the power series are positive, so for $n\geq k_\epsilon$ we have $R^{(n)}(p)\geq (p_c +\epsilon)^{-(n-1)}p^{n-1}$. Note that
 
    \begin{align*}
 
    	R^{(n)}(p)\geq(p_c+\epsilon)^{-(n-1)}p^{n-1}=\left(\frac{p}{p_c+\epsilon}\right)^{n-1} \geq \left(\frac{p}{p_c}\right)^{\frac{n-1}{2}},
 
    \end{align*}
 
    where the last inequality holds for $\epsilon\leq\sqrt{p_c}(\sqrt{p}-\sqrt{p_c})$.
 
    
 
    \section{Calculating the coefficients $a_k^{(n)}$}
 
    Let $\rho'\in\mathbb{R}[p]^{2^n}$ be a vector of polynomials, and let $\text{rank}(\rho')$ be defined in the following way: 
 
    $$\text{rank}(\rho'):=\min_{b\in\{0,1\}^n}\left( |b|+ \text{maximal } k\in\mathbb{N} \text{ such that } p^k \text{ divides } \rho'_b\right).$$
 
	Clearly for any $\rho'$ we have that $\text{rank}(\rho' M_{(n)})\geq \text{rank}(\rho') + 1$. Another observation is, that all elements of $\rho'$ are divisible by $p^{\text{rank}(\rho')-n}$.
 
    We observe that for the initial $\rho$ we have that $\text{rank}(\rho)=n$, therefore $\text{rank}(\rho*(M_{(n)}^k))\geq n+k$, and so $\rho*(M_{(n)}^k)*\mathbbm{1}$ is obviously divisible by $p^{k}$. This implies that $a_k^{(n)}$ can be calculated by only looking at $\rho*(M_{(n)}^1)*\mathbbm{1}, \ldots, \rho*(M_{(n)}^k)*\mathbbm{1}$.
 
    
 
\newpage
 
\section{Quasiprobability method}
 
Let us first introduce notation for paths of the Markov Chain
 
\begin{definition}[Paths]
 
    We define a \emph{path} of the Markov Chain as a sequence of states and resampling choices $\xi=((b_0,r_0),(b_1,r_1),...,(b_k,r_k)) \in (\{0,1\}^n\times[n])^k$ indicating that at time $t$ Markov Chain was in state $b_t\in\{0,1\}^n$ and then resampled site $r_t$. We denote by $|\xi|$ the length $k$ of such a path, i.e. the number of resamples that happened, and by $\mathbb{P}[\xi]$ the probability associated to this path.
 
    We denote by $\paths{b}$ the set of all valid paths $\xi$ that start in state $b$ and end in state $\mathbf{1} := 1^n$.
 
\end{definition}
 
We can write the expected number of resamplings per site $R^{(n)}(p)$ as
 
\begin{align}
 
    R^{(n)}(p) &= \frac{1}{n}\sum_{b\in\{0,1\}^{n}} \rho_b \; R_b(p) \label{eq:originalsum} ,
 
\end{align}
 
where $R_b(p)$ is the expected number of resamplings when starting from configuration $b$
 
\begin{align*}
 
	R_b(p) &= \sum_{\xi \in \paths{b}} \mathbb{P}[\xi] \cdot |\xi| .
 
\end{align*}
 

	
 
We consider $R^{(n)}(p)$ as a power series in $p$ and show that many terms in (\ref{eq:originalsum}) cancel out if we only consider the series up to some finite order $p^k$. The main idea is that if a path samples a $0$ then $\mathbb{P}[\xi]$ gains a factor $p$ so paths that contribute to $p^k$ can't be arbitrarily long.\\
 

	
 
To see this, we split the sum in (\ref{eq:originalsum}) into parts that will later cancel out. The initial probabilities $\rho_b$ contain a factor $p$ for every $0$ and a factor $(1-p)$ for every $1$. When expanding this product of $p$s and $(1-p)$s, we see that the $1$s contribute a factor $1$ and a factor $(-p)$ and the $0$s only give a factor $p$. We want to expand this product explicitly and therefore we no longer consider bitstrings $b\in\{0,1\}^n$ but bitstrings $b\in\{0,1,1'\}^n$. We view this as follows: every site can have one of $\{0,1,1'\}$ with `probabilities' $p$, $1$ and $-p$ respectively. A configuration $b=101'1'101'$ now has probability $\rho_{b} = 1\cdot p\cdot(-p)\cdot(-p)\cdot 1\cdot p\cdot(-p) = -p^5$ in the starting state $\rho$. It should not be hard to see that we have
 
\begin{align*}
 
    R^{(n)}(p) &= \frac{1}{n}\sum_{b\in\{0,1,1'\}^{n}} \rho_{b} \; R_{\bar{b}}(p) ,
 
\end{align*}
 
where $\bar{b}$ is the bitstring obtained by changing every $1'$ in it back to a $1$. It is simply the same sum as (\ref{eq:originalsum}) but now every factor $(1-p)$ is explicitly split into $1$ and $(-p)$.
 
   
 
Some terminology: for any configuration we call a $0$ a \emph{particle} (probability $p$) and a $1'$ an \emph{antiparticle} (probability $-p$). We use the word \emph{slot} for a position that is occupied by either a paritcle or antiparticle ($0$ or $1'$). In the initial state, the probability of a configuration is given by $\pm p^{\mathrm{\#slots}}$ where the $\pm$ sign depends on the parity of the number of antiparticles.
 
    
 
We can further rewrite the sum over $b\in\{0,1,1'\}^n$ as a sum over all slot configurations $C\subseteq[n]$ and over all possible fillings of these slots.
 
\begin{align*}
 
	R^{(n)}(p) &= \frac{1}{n} \sum_{C\subseteq[n]} \sum_{f\in\{0,1'\}^{|C|}} \rho_{C(f)} R_{C(f)} ,
 
\end{align*}
 
where $C(f)\in\{0,1,1'\}^n$ denotes a configuration with slots on the sites $C$ filled with (anti)particles described by $f$. The non-slot positions are filled with $1$s.
 

	
 
\begin{definition}[Diameter and gaps] \label{def:diameter} \label{def:gaps}
 
    For a subset $C\subseteq[n]$, we define the \emph{diameter} $\diam{C}$ to be the minimum size of an interval $I$ containing $C$. Here we consider both $C$ and the interval modulo $n$. In other words $\diam{C} = \min\{ j \vert \exists i : C\subseteq [i,i+j-1] \}$. We define the \emph{gaps} of $C$, as $I\setminus C$ and denote this by $\gaps{C}$. Note that $\diam{C} = |C| + |\gaps{C}|$.  Define $\maxgap{C}$ as the size of the largest connected component of $\gaps{C}$. Figure \ref{fig:diametergap} illustrates these concepts with a picture. 
 
\end{definition}
 
\begin{figure}
 
	\begin{center}
 
    	\includegraphics{diagram_gap.pdf}
 
    \end{center}
 
    \caption{\label{fig:diametergap} Illustration of Definition \ref{def:diameter}. A set $C=\{1,2,4,7,9\}\subseteq[n]$ consisting of 5 positions is shown by the red dots. The smallest interval containing $C$ is $[1,9]$, so the diameter is $\diam{C}=9$. The blue squares denote the set $\gaps{C} = \{3,5,6,8\}$. The dotted line at the top depicts the rest of the circle which may be much larger. The largest gap of $C$ is $\maxgap{C}=2$ which is the largest connected component of $\gaps{C}$.}
 
\end{figure}
 

	
 
\begin{claim}[Strong cancellation claim] \label{claim:strongcancel}
 
	The lowest order term in
 
    \begin{align*}
 
        \sum_{f\in\{0,1'\}^{|C|}} \rho_{C(f)} R_{C(f)} ,
 
    \end{align*}
 
	is $p^{\diam{C}}$ when $n$ is large enough. All lower order terms cancel out.
 
\end{claim}
 

	
 
Example: for $C_0=\{1,2,4,7,9\}$ (the configuration shown in Figure \ref{fig:diametergap}) we computed the quantity up to order $p^{20}$ in an infinite system:
 
\begin{align*}
 
	\sum_{f\in\{0,1'\}^{|C_0|}} \rho_{C_0(f)} R_{C_0(f)} &= 0.0240278 p^{9} + 0.235129 p^{10} + 1.24067 p^{11} + 4.71825 p^{12} \\
 
    &\quad + 14.5555 p^{13} + 38.8307 p^{14} + 93.2179 p^{15} + 206.837 p^{16}\\
 
    &\quad + 432.302 p^{17} + 862.926 p^{18} + 1662.05 p^{19} + 3112.9 p^{20} + \mathcal{O}(p^{21})
 
\end{align*}
 
and indeed the lowest order is $\diam{C}=9$.
 

	
 
~
 

	
 
A weaker version of the claim is that if $C$ contains a gap of size $k$, then the sum is zero up to and including order $p^{|C|+k-1}$.
 
\begin{claim}[Weak cancellation claim] \label{claim:weakcancel}
 
	For $C\subseteq[n]$ a configuration of slot positions, the lowest order term in
 
    \begin{align*}
 
        \sum_{f\in\{0,1'\}^{|C|}} \rho_{C(f)} R_{C(f)} ,
 
    \end{align*}
 
    is at least $p^{|C|+\maxgap{C}}$ when $n$ is large enough. All lower order terms cancel out.
 
\end{claim}
 
This weaker version would imply \ref{it:const} but for $\mathcal{O}(k^2)$ as opposed to $k+1$.
 

	
 
\newpage
 
The reason that claim \ref{claim:strongcancel} would prove \ref{it:const} is the following: to know the value of $a_k^{(n)}$, for any $n\geq k+1$ it is enough to look at configurations $C$ with diameter at most $k$, since larger configurations do not contribute to $a_k^{(n)}$.
 
For a starting state $b\in\{0,1\}^n$ that \emph{does} give a nonzero contribution, you can take that same starting configuration and translate it to get $n$ other configurations that give the same contribution. (An exception is a starting state like $1010101010...$ which you can only translate twice, but we only have to consider configurations with small diameter, in which case you can make exactly $n$ translations.)
 
Therefore the coefficient in the expected number of resamplings is a multiple of $n$ which Andr\'as already divided out in the definition of $R^{(n)}(p)$. To show \ref{it:const} we argue that this is the \emph{only} dependency on $n$. This is because there are only finitely many (depending on $k$ but not on $n$) configurations where the $k$ slots are nearby regardless of the value of $n$. So there are only finitely many nonzero contributions after translation symmetry was taken out. For example, when considering all starting configurations with 5 slots one might think there are $\binom{n}{5}$ configurations to consider which would be a dependency on $n$ (more than only the translation symmetry). But since most of these configurations have a diameter larger than $k$, they do not contribute to $a_k$. Only finitely many do and that does not depend on $n$.
 

	
 
~
 

	
 
Section \ref{sec:computerb} shows how to compute $R_b$ (this is not relevant for showing the claim) and the section after that shows how to prove the weaker claim.
 

	
 
\newpage
 
\subsection{Computation of $R_b$} \label{sec:computerb}
 

	
 
By $R_{101}$ we denote $R_b(p)$ for a $b$ that consists of only $1$s except for a single zero. We compute $R_{101}$ up to second order in $p$. This requires the following transitions.
 
\begin{align*}
 
    \framebox{$1 0 1$} &\to \framebox{$1 1 1$} & (1-p)^3 = 1-3p+3p^2-p^3\\
 
    \hline
 
    \framebox{$1 0 1$} &\to
 
        \begin{cases}
 
            \framebox{$0 1 1$}\\
 
            \framebox{$1 0 1$}\\
 
            \framebox{$1 1 0$}
 
        \end{cases}
 
        & 3p(1-p)^2 = 3p-6p^2+3p^3\\
 
    \hline
 
    \framebox{$1 0 1$} &\to \framebox{$0 1 0$} & p^2(1-p) = p^2-p^3\\
 
    \framebox{$1 0 1$} &\to
 
        \begin{cases}
 
            \framebox{$1 0 0$}\\
 
            \framebox{$0 0 1$}
 
        \end{cases}
 
        & 2p^2(1-p) = 2p^2 - 2p^3\\
 
    \hline
 
    \framebox{$1 0 1$} &\to \framebox{$0 0 0$} & p^3
 
\end{align*}
 
With this we can write a recursive formula for the expected number of resamples from $101$:
 
\begin{align*}
 
    R_{101} &= (1-3p+3p^2 - p^3)(1) + (3p -6p^2 +3p^3) (1+R_{101}) \\
 
            &\quad + (p^2 - p^3) (1+R_{10101}) + (2p^2-2p^3) (1+R_{1001}) \\
 
			&= 1 + 3 p + 7 p^2 + 14.6667 p^3 + 29 p^4 + 55.2222 p^5 + 102.444 p^6 + 186.36 p^7 \\
 
            &\quad + 333.906 p^8 + 590.997 p^9 + 1035.58 p^{10} + 1799.39 p^{11} + 3104.2 p^{12} \\
 
            &\quad+ 5322.18 p^{13} + 9075.83 p^{14} + 15403.6 p^{15} + 26033.4 p^{16} + 43833.5 p^{17} \\
 
            &\quad+ 73555.2 p^{18} + 123053 p^{19} + 205290 p^{20} + 341620 p^{21} + 567161 p^{22} \\
 
            &\quad+ 939693 p^{23} + 1.5537\cdot10^{6} p^{24} + 2.56158\cdot10^{6} p^{25} + \mathcal{O}(p^{26})
 
\end{align*}
 
where the recursion steps were done with a computer for an infinite line (or a cirlce where $n$ is assumed to be much larger than the largest power of $p$ considered).
 

	
 
Note: in the first line at the second term it uses that with probability $(3p-6p^2 + 3p^3)$ the state goes to $\framebox{$101$}$ and then the expected number of resamplings is $1+R_{101}$. Note that the actual term in the recursive formula should be
 
$$(3p-6p^2+3p^3)\cdot\left( \sum_{\xi\in\paths{101}} \mathbb{P}[\xi] \cdot \left( 1 + |\xi|\right) \right) = (3p-6p^2+3p^3)\left( p_\mathrm{tot} + R_{101} \right)$$
 
where $p_\mathrm{tot} := \sum_{\xi\in\paths{b}} \mathbb{P}[\xi]$. However, since the state space is finite (for finite $n$) and there is always a non-vanishing probability to go to $\mathbf{1}$, we know that $p_\mathrm{tot}=1$, i.e. the process terminates almost surely.
 

	
 
\newpage
 
\subsection{Weak cancellation proof}
 

	
0 comments (0 inline, 0 general)