\documentclass[orivec]{llncs} \bibliographystyle{splncs} \let\proof\relax \let\endproof\relax \usepackage{amsmath,amssymb,amsthm,bbm} \usepackage{mathpartir} \usepackage{stmaryrd} \usepackage{soul} \usepackage{listings} \usepackage{proof} \usepackage{color} \usepackage{hyperref} \usepackage{float} \usepackage{rotating} \usepackage{mdframed} \usepackage{enumitem} \usepackage{xifthen} \usepackage{tikz} \usepackage{multirow} \usepackage[misc]{ifsym} \addtolength{\belowcaptionskip}{-0.5em} \newif\ifdraft \draftfalse \newcommand{\esyntax}[1]{\mathsf{#1}} \newcommand{\rname}[1]{[\textsf{#1}]} \newcommand{\jrhol}[7]{#1 \mid #2 \vdash #3 : #4 \sim #5 : #6 \mid #7} \newcommand{\jrholu}[5]{#1 \mid #2 \vdash #3 \sim #4 \mid #5} \newcommand{\sjrhol}[7]{#2 \vdash #3 \sim #5 \mid #7} \newcommand{\juhol}[5]{#1 \mid #2 \vdash #3 : #4 \mid #5} \newcommand{\sjuhol}[5]{#2 \vdash #3 \mid #5} \newcommand{\jhol}[3]{#1 \mid #2 \vdash #3} \newcommand{\vhol}[3]{#1;#2 \models #3} \newcommand{\jholn}[3]{#1 \mid #2 \vdash_{\sf HOL} #3} \newcommand{\jlc}[3]{#1 \vdash #2 : #3} \newcommand{\jwf}[2]{#1 \vdash #2 \; \sf{wf}} \newcommand{\subst}[2]{[#2/ #1]} \newcommand{\ltag}{_1} \newcommand{\rtag}{_2} \newcommand{\res}{\mathbf{r}} \newcommand{\sem}[1]{\llbracket #1 \rrbracket} \newcommand{\semv}[2]{\llbracket #1 \rrbracket_{#2} } \newcommand{\semexp}[1]{\llparenthesis #1 \rrparenthesis} \newcommand{\semexpv}[2]{\llparenthesis #1 \rrparenthesis_{#2} } \newcommand{\pair}[2]{\langle #1, #2 \rangle} \newcommand{\nat}{\mathbb{N}} \newcommand{\zint}{\mathbb{Z}} \newcommand{\real}{\mathbb{R}} \newcommand{\unit}{\mathbb{U}} \newcommand{\listt}[1]{\mathsf{list}_{#1}} \newcommand{\bool}{\mathbb{B}} \newcommand{\tbool}{\mathsf{tt}} \newcommand{\fbool}{\mathsf{ff}} \newcommand{\nil}{[]} \newcommand{\cons}[2]{#1\! ::\! #2} \newcommand{\casebool}[3]{{\rm case}\ #1\ {\sf of}\ \tbool \mapsto #2 ; \fbool \mapsto #3} \newcommand{\casenat}[3]{{\rm case}\ #1\ {\sf of}\ 0 \mapsto #2 ; S \mapsto #3} \newcommand{\caselist}[3]{{\rm case}\ #1\ {\rm of}\ \nil \mapsto #2 ; \cons{\_}{\_} \mapsto #3} \newcommand{\defsubst}[2]{\subst{\res\ltag}{#1}\subst{\res\rtag}{#2}} \newcommand{\wdef}[3]{\mathcal{D}\hspace{-0.2mm}\mathit{ef}(#1,#2,#3)} \newcommand{\letrec}[3]{{\rm letrec}\ {\mathit #1}\ #2 = #3} \newcommand{\assn}{\star} \newcommand{\cost}[1]{\llparenthesis #1 \rrparenthesis} \newcommand{\costt}[1]{\lfloor #1 \rfloor} \newcommand{\rcost}[1]{\| #1 \|} \newcommand{\rcostt}[1]{\llfloor #1 \rrfloor} \newcommand{\defeq}{\triangleq} \newcommand{\erase}[1]{| #1 |} \newcommand{\reftype}[3]{\ifthenelse{\isempty{#1}}{\{\res : #2 \mid #3\}}{\{#1 : #2 \mid #3\}}} \newcommand{\ertype}[2]{\ifthenelse{\isempty{#2}}{\lfloor #1 \rfloor}{\lfloor #1 \rfloor(#2)}} \newcommand{\errtype}[2]{\ifthenelse{\isempty{#2}}{\llfloor #1 \rrfloor}{\llfloor #1 \rrfloor(#2)}} \newcommand{\erfor}[1]{| #1 |} \newcommand{\embind}[1]{[#1]} \newcommand{\emfor}[2]{\ifthenelse{\isempty{#2}}{#1}{#1\subst{\res}{#2}}} \newcommand{\emterm}[2]{\ifthenelse{\isempty{#2}}{#1}{#1\subst{\res}{#2}}} \newcommand{\emtyj}[3]{\juhol{\erfor{#1}}{\ertype{#1}{}}{#2}{\erfor{#3}}{\ertype{#3}{\res}}} \newcommand{\jreft}[3]{#1 \vdash #2 : #3} \newcommand{\pitype}[3]{\Pi(#1:#2).#3} \newcommand{\ttmonad}[1]{\mathbb{T}_{#1}} \newcommand{\eret}[1]{\eta_{#1}} \newcommand{\ebind}{\mathsf{bind}} \newcommand{\dccform}[2]{\lfloor #2 \rfloor_{#1}} \newcommand{\adversary}{a} \newcommand{\boolform}[1]{R^{\bool}_{#1}} \newcommand{\wexec}{\mbox{\scriptsize exec}} \newcommand{\wdiff}{\mbox{\scriptsize diff}} \newcommand{\ldiff}{\lesssim} \newcommand{\uarr}[2]{\mathrel{\xrightarrow[]{\wexec(#1,#2)}}} \newcommand{\tforall}[3]{\forall#1\overset{\wexec(#2,#3)}{::}S.\,} \newcommand{\texists}[1]{\exists#1{::}S.\,} \newcommand{\tarrd}[1]{\mathrel{\xrightarrow{\wdiff(#1)}}} \newcommand{\tforalld}[2]{\forall#1\overset{\wdiff(#2)}{::}S.\,} \newcommand{\tbox}[1]{\square\,#1} \newcommand{\jtype}[4]{\mathrel{\vdash_{#1}^{#2} {#3} : #4}} \newcommand{\jtypediff}[4]{\mathrel{\vdash {#2} \ominus {#3} \ldiff #1 : {#4}}} \newcommand{\econs}{\mbox{cons}} \newcommand{\ulist}[2]{\mbox{list}_{#2}[#1]} \newcommand{\ctx}{\Delta; \Phi_a; \Gamma} \newcommand{\kw}[1]{\mathtt{#1}} \newcommand{\econst}{\kw{n}} \newcommand{\trint}{\mbox{int}_r} \newcommand{\tlist}[3]{\mbox{list}[#1]^{#2}\,#3} \newcommand{\ecase}{\mbox{\;case\;}} \newcommand{\eof}{\mbox{\;of\;}} \newcommand{\eas}{\mbox{\;as\;}} \newcommand{\ecelim}{\mbox{celim\;}} \newcommand{\enil}{\mbox{nil\;}} \newcommand{\fiv}[1]{\text{FIV}(#1)} \newcommand{\eLam}{ \Lambda} \newcommand{\elam}{ \lambda} \newcommand{\eApp}{ [\,]\,} \newcommand{\sat}[1]{\models#1} \newcommand{\jsubtype}[2]{\sat#1\sqsubseteq#2} \newcommand{\trmo}[1]{|#1|} \newcommand{\grt}{A} \newcommand{\tcho}[1]{U\,#1} \newcommand{\octx}{\Delta; \Phi_a; \Omega} \newcommand{\tint}{\mbox{int}} \newcommand{\kapp}{c_{app}} \def\tyle{\preceq} \ifdraft \newcommand{\aag}[1]{\textit{\color{blue}[AA]: #1}} \newcommand{\gb}[1]{\textit{\color{red}[GB]: #1}} \newcommand{\mg}[1]{\textit{\color{cyan}[MG]: #1}} \newcommand{\dg}[1]{\textit{\color{brown}[DG]: #1}} \newcommand{\ales}[1]{\textit{\color{green}[AB]: #1}} \newcommand{\lb}[1]{\textit{\color{green}[LB]: #1}} \else \newcommand{\aag}[1]{} \newcommand{\gb}[1]{} \newcommand{\mg}[1]{} \newcommand{\dg}[1]{} \newcommand{\ales}[1]{} \newcommand{\lb}[1]{} \fi \theoremstyle{definition} \newtheorem{thm}{Theorem} \newtheorem{lem}[thm]{Lemma} \newtheorem{cor}[thm]{Corollary} \newtheorem{prop}[thm]{Proposition} \newtheorem{eg}[thm]{Example} \newtheorem{hypothesis}[thm]{Hypothesis} \newtheorem{motivation}{Motivation} \newtheorem{defn}[thm]{Definition} \newcommand{\RHOL}{\mathsf{2HOL}} \newcommand{\renv}[1]{\mathcal{#1}} \newcommand{\vbar}[0]{\mathrel{|}} \newcommand{\rtprod}[3]{\Pi ({#1} :: {#2}) .\, {#3}} \newcommand{\rtref}[3]{\{ {#1} :: {#2} \vbar {#3} \}} \newcommand{\tprod}[3]{\Pi ({#1} : {#2}) .\, {#3}} \newcommand{\tref}[3]{\{ {#1} : {#2} \vbar {#3} \}} \newcommand{\tyinterp}[1]{\llbracket {#1} \rrbracket} \newcommand{\interp}[2]{\llbracket {#2} \rrbracket_{#1}} \newcommand{\rinterp}[2]{\llparenthesis {#2} \rrparenthesis_{#1}} \DeclareDocumentCommand{\later}{ o m}{ \IfNoValueTF{#1} {\triangleright #2} {\triangleright\hrt{#1} . #2}} \DeclareDocumentCommand{\latern}{ o m}{ \IfNoValueTF{#1} {\triangleright #2} {\triangleright #1 . #2}} \newcommand{\vrt}[1]{\left[\begin{array}{l} #1 \end{array}\right]} \newcommand{\hrt}[1]{\left[ #1 \right]} \newcommand{\dsubst}[3]{\triangleright[#1 \leftarrow #2].#3} \newcommand{\nextt}[2]{\triangleright[#1].#2} \newcommand{\prev}[1]{{\rm prev}\ #1} \newcommand{\boxx}[1]{{\rm box}\ #1} \newcommand{\jguhol}[7]{#1 \mid #2 \mid #3 \mid #4 \vdash #5 : #6 \mid #7} \newcommand{\jguholsc}[7]{\Omega \vdash #5 : #6 \mid #7} \newcommand{\letbox}[3]{{\rm letb}\ #1 \leftarrow #2\ {\rm in}\ #3} \newcommand{\unbox}[1]{{\rm unbox}\ #1} \newcommand{\letconst}[3]{{\rm letc}\ #1 \leftarrow #2\ {\rm in}\ #3} \newcommand{\conshat}[2]{#1\hat{::}#2} \newcommand{\fix}[2]{{\rm fix}\ #1.\ #2} \newcommand{\rec}[2]{{\rm rec}\ #1.\ #2} \newcommand{\app}{\circledast} \newcommand{\jgrhol}[9]{#1 \mid #2 \mid #3 \mid #4 \vdash #5 : #6 \sim #7 : #8 \mid #9} \newcommand{\jgrholsc}[9]{\Omega \vdash #5 : #6 \sim #7 : #8 \mid #9} \newcommand{\jgrholnoc}[5]{\vdash #1 : #2 \sim #3 : #4 \mid #5} \newcommand{\jgrholnocnot}[3]{\vdash #1 \sim #2 \mid #3} \newcommand{\jghol}[5]{#1 \mid #2 \mid #3 \mid #4 \vdash #5} \newcommand{\jgholsc}[5]{\Omega \vdash #5} \newcommand{\conj}{\mathrel{\wedge}} \newcommand{\trees}{\mathcal{S}} \newcommand{\ot}{\leftarrow} \newcommand{\xot}[1]{\xleftarrow{#1}} \newcommand{\xto}[1]{\xrightarrow{#1}} \newcommand{\To}{\Rightarrow} \newcommand{\oT}{\Leftarrow} \newcommand{\ii}{\mathit} \newcommand{\thesystem}{gRHOL} \newcommand{\inj}[1]{{\rm inj}_{#1}} \newcommand{\CASE}{\operatorname{\mathrm{case}}} \newcommand{\OF}{\operatorname{\mathrm{of}}} \newcommand{\IN}{\operatorname{\mathrm{in}}} \newcommand{\hd}{\operatorname{\mathrm{hd}}} \newcommand{\tl}{\operatorname{\mathrm{tl}}} \newcommand{\Distr}{\mathsf{D}} \newcommand{\Str}[1]{\ifstrempty{#1}{\operatorname{\mathrm{Str}}}{\operatorname{\mathrm{Str}}_{#1}}} \newcommand{\munit}[1]{\mathsf{munit}(#1)} \newcommand{\mlet}[3]{\mathsf{let}~#1=#2~\mathsf{in}~#3} \newcommand{\carac}[1]{\mathbbm{1}_{#1}} \newcommand{\supp}{\mathsf{supp}} \newcommand{\conslist}[2]{#1 :: #2} \newcommand{\emptylist}{\operatorname{[]}} \newcommand{\fold}{\operatorname{\mathsf{fold}}} \newcommand{\unfold}{\operatorname{\mathsf{unfold}}} \newcommand{\markov}{\operatorname{\mathsf{markov}}} \newcommand{\isetsep}{\;\ifnum\currentgrouptype=16 \middle\fi|\;} \newcommand{\bern}[1]{\mathcal{B}(#1)} \newcommand{\unif}[1]{\mathcal{U}_{#1}} \newcommand{\rfrac}[2]{{}^{#1}\!/\!_{#2}} \newcommand{\All}{\operatorname{All}} \newcommand{\coupling}[3]{#2 \mathrel{\mathcal{L}\left(#1\right)} #3} \newcommand{\coupl}[3]{\diamond_{#1, #2}. #3} \newcommand{\sP}[2]{\operatorname{Prob}_{#1}[#2]} \newcommand{\shrinkcaption}{\vspace{-1em}} \begin{document} \title{Relational Reasoning for Markov Chains in a Probabilistic Guarded Lambda Calculus} \author{Alejandro Aguirre\inst{1}$^\text{(\Letter)}$ \and Gilles Barthe\inst{1} \and Lars Birkedal\inst{2} \and Ale\u s Bizjak\inst{2} \and \\ Marco Gaboardi\inst{3} \and Deepak Garg\inst{4}} \institute{IMDEA Software Institute \\ \texttt{alejandro.aguirre@imdea.org} \and Aarhus University \and University at Buffalo, SUNY \and MPI-SWS} \maketitle \begin{abstract} We extend the simply-typed guarded $\lambda$-calculus with discrete probabilities and endow it with a program logic for reasoning about relational properties of guarded probabilistic computations. This provides a framework for programming and reasoning about infinite stochastic processes like Markov chains. We demonstrate the logic sound by interpreting its judgements in the topos of trees and by using probabilistic couplings for the semantics of relational assertions over distributions on discrete types. The program logic is designed to support syntax-directed proofs in the style of relational refinement types, but retains the expressiveness of higher-order logic extended with discrete distributions, and the ability to reason relationally about expressions that have different types or syntactic structure. In addition, our proof system leverages a well-known theorem from the coupling literature to justify better proof rules for relational reasoning about probabilistic expressions. We illustrate these benefits with a broad range of examples that were beyond the scope of previous systems, including shift couplings and lump couplings between random walks. \end{abstract} \section{Introduction} Stochastic processes are often used in mathematics, physics, biology or finance to model evolution of systems with uncertainty. In particular, Markov chains are ``memoryless'' stochastic processes, in the sense that the evolution of the system depends only on the current state and not on its history. Perhaps the most emblematic example of a (discrete time) Markov chain is the simple random walk over the integers, that starts at 0, and that on each step moves one position either left or right with uniform probability. Let $p_i$ be the position at time $i$. Then, this Markov chain can be described as: \[ p_0 = 0 \quad\quad p_{i+1} = \begin{cases} p_i + 1\ \text{with probability}\ 1/2 \\ p_i - 1\ \text{with probability}\ 1/2 \end{cases} \] The goal of this paper is to develop a programming and reasoning framework for probabilistic computations over infinite objects, such as Markov chains. Although programming and reasoning frameworks for infinite objects and probabilistic computations are well-understood in isolation, their combination is challenging. In particular, one must develop a proof system that is powerful enough for proving interesting properties of probabilistic computations over infinite objects, and practical enough to support effective verification of these properties. \paragraph*{Modelling probabilistic infinite objects} A first challenge is to model probabilistic infinite objects. We focus on the case of Markov chains, due to its importance. A (discrete-time) Markov chain is a sequence of random variables $\{X_i\}$ over some fixed type $T$ satisfying some independence property. Thus, the straightforward way of modelling a Markov chain is as a \emph{stream of distributions} over $T$. Going back to the simple example outlined above, it is natural to think about this kind of \emph{discrete-time} Markov chain as characterized by the sequence of positions $\{p_i\}_{i \in \nat}$, which in turn can be described as an infinite set indexed by the natural numbers. This suggests that a natural way to model such a Markov chain is to use \emph{streams} in which each element is produced \emph{probabilistically} from the previous one. However, there are some downsides to this representation. First of all, it requires explicit reasoning about probabilistic dependency, since $X_{i+1}$ depends on $X_i$. Also, we might be interested in global properties of the executions of the Markov chain, such as ``The probability of passing through the initial state infinitely many times is 1''. These properties are naturally expressed as properties of the whole stream. For these reasons, we want to represent Markov chains as \emph{distributions over streams}. Seemingly, one downside of this representation is that the set of streams is not countable, which suggests the need for introducing heavy measure-theoretic machinery in the semantics of the programming language, even when the underlying type is discrete or finite. Fortunately, measure-theoretic machinery can be avoided (for discrete distributions) by developing a probabilistic extension of the simply-typed guarded $\lambda$-calculus and giving a semantic interpretation in the topos of trees~\cite{CBGB16}. Informally, the simply-typed guarded $\lambda$-calculus~\cite{CBGB16} extends the simply-typed lambda calculus with a \emph{later} modality, denoted by $\later$. The type $\later{A}$ ascribes expressions that are available one unit of logical time in the future. The $\later$ modality allows one to model infinite types by using \lq\lq finite\rq\rq\ approximations. For example, a stream of natural numbers is represented by the sequence of its (increasing) prefixes in the topos of trees. The prefix containing the first $i$ elements has the type $S_i \defeq \nat \times \later{\nat} \times \ldots \times \later^{(i-1)}{\nat}$, representing that the first element is available now, the second element a unit time in the future, and so on. This is the key to representing probability distributions over infinite objects without measure-theoretic semantics: We model probability distributions over non-discrete sets as discrete distributions over their (the sets') approximations. For example, a distribution over streams of natural numbers (which a priori would be non-discrete since the set of streams is uncountable) would be modelled by a \emph{sequence of distributions} over the finite approximations $S_1, S_2, \ldots$ of streams. Importantly, since each $S_i$ is countable, each of these distributions can be discrete. \dg{As Alejandro also noted after his example that I commented out, it's unclear what's so specific or special about the $\later$ modality. It seems that we could use \emph{any} approximating sequence of discrete sets to avoid the issue with measure-theory. What's special about $\later$ and guards in our context?} \lb{the use of $\later$ and guards just allow us to use types to ensure productivity (and contractiveness to ensure existence of fixed points)} \mg{My understanding is that there is nothing special but we need in the logic to have a way to talk about these approximations, and $\later$ and guards gives us a natural way to do this. no?} \paragraph*{Reasoning about probabilistic computations} Probabilistic computations exhibit a rich set of properties. One natural class of properties is related to probabilities of events, saying, for instance, that the probability of some event $E$ (or of an indexed family of events) increases at every iteration. However, several interesting properties of probabilistic computation, such as stochastic dominance or convergence (defined below) are relational, in the sense that they refer to two runs of two processes. In principle, both classes of properties can be proved using a higher-order logic for probabilistic expressions, e.g.\, the internal logic of the topos of trees, suitably extended with an axiomatization of finite distributions. However, we contend that an alternative approach inspired from refinement types is desirable and provides better support for effective verification. More specifically, reasoning in a higher-order logic, e.g.\, in the internal logic of the topos of trees, does not exploit the \emph{structure of programs} for non-relational reasoning, nor the \emph{structural similarities} between programs for relational reasoning. As a consequence, reasoning is more involved. To address this issue, we define a relational proof system that exploits the structure of the expressions and supports syntax-directed proofs, with necessary provisions for escaping the syntax-directed discipline when the expressions do not have the same structure. The proof system manipulates judgements of the form: \begin{equation*} \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi} \label{eq:judgement-2} \end{equation*} where $\Delta$ and $\Gamma$ are two typing contexts, $\Sigma$ and $\Psi$ respectively denote sets of assertions over variables in these two contexts, $t_1$ and $t_2$ are well-typed expressions of type $A_1$ and $A_2$, and $\phi$ is an assertion that may contain the special variables $\res\ltag$ and $\res\rtag$ that respectively correspond to the values of $t_1$ and $t_2$. The context $\Delta$ and $\Gamma$, the terms $t_1$ and $t_2$ and the types $A_1$ and $A_2$ provide a specification, while $\Sigma$, $\Psi$, and $\phi$ are useful for reasoning about relational properties over $t_1,t_2$, their inputs and their outputs. This form of judgement is similar to that of Relational Higher-Order Logic \cite{ABGGS17}, from which our system draws inspiration. In more detail, our relational logic comes with typing rules that allow one to reason about relational properties by exploiting as much as possible the syntactic similarities between $t_1$ and $t_2$, and to fall back on pure logical reasoning when these are not available. In order to apply relational reasoning to guarded computations the logic provides relational rules for the later modality $\later{}$ and for a related modality $\square{}$, called ``constant''. These rules allow the relational verification of general relational properties that go beyond the traditional notion of program equivalence and, moreover, they allow the verification of properties of guarded computations over different types. The ability to reason about computations of different types provides significant benefits over alternative formalisms for relational reasoning. For example, it enables reasoning about relations between programs working on different data structures, e.g. a relation between a program working on a stream of natural numbers, and a program working on a stream of pairs of natural numbers, or having different structures, e.g. a relation between an application and a case expression. Importantly, our approach for reasoning formally about probabilistic computations is based on \emph{probabilistic couplings}, a standard tool from the analysis of Markov chains~\cite{Lindvall02,Thorisson00}. From a verification perspective, probabilistic couplings go beyond equivalence properties of probabilistic programs, which have been studied extensively in the verification literature, and yet support compositional reasoning~\cite{BartheEGHSS15,BartheGHS17}. The main attractive feature of coupling-based reasoning is that it limits the need of explicitly reasoning about the probabilities---this avoids complex verification conditions. We provide sound proof rules for reasoning about probabilistic couplings. Our rules make several improvements over prior relational verification logics based on couplings. First, we support reasoning over probabilistic processes of different types. Second, we use Strassen's theorem~\cite{strassen1965existence} a remarkable result about probabilistic couplings, to achieve greater expressivity. Previous systems required to prove a bijection between the sampling spaces to show the existence of a coupling~\cite{BartheEGHSS15,BartheGHS17}, Strassen's theorem gives a way to show their existence which is applicable in settings where the bijection-based approach cannot be applied. And third, we support reasoning with what are called shift couplings, coupling which permits to relate the states of two Markov chains at possibly different times (more explanations below). \paragraph*{Case studies} We show the flexibility of our formalism by verifying several examples of relational properties of probabilistic computations, and Markov chains in particular. These examples cannot be verified with existing approaches. First, we verify a classic example of probabilistic non-interference which requires the reasoning about computations at different types. Second, in the context of Markov chains, we verify an example about stochastic dominance which exercises our more general rule for proving the existence of couplings modelled by expressions of different types. Finally, we verify an example involving shift relations in an infinite computation. This style of reasoning is motivated by \lq\lq shift\rq\rq\ couplings in Markov chains. In contrast to a standard coupling, which relates the states of two Markov chains at the same time $t$, a shift coupling relates the states of two Markov chains at possibly different times. Our specific example relates a standard random walk (described earlier) to a variant called a lazy random walk; the verification requires relating the state of standard random walk at time $t$ to the state of the lazy random walk at time $2t$. We note that this kind of reasoning is impossible with conventional relational proof rules even in a non-probabilistic setting. Therefore, we provide a novel family of proof rules for reasoning about shift relations. At a high level, the rules combine a careful treatment of the later and constant modalities with a refined treatment of fixpoint operators, allowing us to relate different iterates of function bodies. \subsection*{Summary of contributions} With the aim of providing a general framework for programming and reasoning about Markov chains, the three main contributions of this work are: \begin{enumerate} \item A probabilistic extension of the guarded $\lambda$-calculus, that enables the definition of Markov chains as discrete probability distributions over streams. \item A relational logic based on coupling to reason in a syntax-directed manner about (relational) properties of Markov chains. This logic supports reasoning about programs that have different types and structures. Additionally, this logic uses results from the coupling literature to achieve greater expressivity than previous systems. \item An extension of the relational logic that allows to relate the states of two streams at possibly different times. This extension supports reasoning principles, such as shift couplings, that escape conventional relational logics. \end{enumerate} \section{Mathematical preliminaries} This section reviews the definition of discrete probability sub-distributions and introduces mathematical couplings. \begin{definition}[Discrete probability distribution] Let $C$ be a discrete (i.e., finite or countable) set. A (total) distribution over $C$ is a function $\mu : C \to [0,1]$ such that $ \sum_{x\in C} \mu(x) = 1 .$ The support of a distribution $\mu$ is the set of points with non-zero probability, $ \supp\ \mu \defeq \{x \in C \mid \mu(x) > 0 \} .$ We denote the set of distributions over $C$ as $\Distr(C)$. Given a subset $E \subseteq C$, the probability of sampling from $\mu$ a point in $E$ is denoted $\Pr_{x\leftarrow \mu}[x \in E]$, and is equal to $\sum_{x \in E} \mu(x)$. \end{definition} \begin{definition}[Marginals] Let $\mu$ be a distribution over a product space $C_1\times C_2$. The first (second) marginal of $\mu$ is another distribution $\Distr(\pi_1)(\mu)$ ($\Distr(\pi_2)(\mu)$) over $C_1$ ($C_2$) defined as: \[\Distr(\pi_1)(\mu)(x) = \sum_{y \in C_2} \mu(x,y) \qquad \left(\Distr(\pi_2)(\mu)(y) = \sum_{x \in C_1} \mu(x,y) \right) \] \end{definition} \subsubsection{Probabilistic couplings} Probabilistic couplings are a fundamental tool in the analysis of Markov chains. When analyzing a relation between two probability distributions it is sometimes useful to consider instead a distribution over the product space that somehow ``couples'' the randomness in a convenient manner. Consider for instance the case of the following Markov chain, which counts the total amount of tails observed when tossing repeatedly a biased coin with probability of tails $p$: \[ n_0 = 0 \quad\quad n_{i+1} = \left\{\begin{array}{l} n_i + 1\ \text{with probability}\ p \\ n_i\ \text{with probability}\ (1-p) \end{array}\right. \] If we have two biased coins with probabilities of tails $p$ and $q$ with $p\leq q$ and we respectively observe $\{n_i\}$ and $\{m_i\}$ we would expect that, in some sense, $n_i \leq m_i$ should hold for all $i$ (this property is known as stochastic dominance). A formal proof of this fact using elementary tools from probability theory would require to compute the cumulative distribution functions for $n_i$ and $m_i$ and then to compare them. The coupling method reduces this proof to showing a way to pair the coin flips so that if the first coin shows tails, so does the second coin. We now review the definition of couplings and state relevant properties. \begin{definition}[Couplings] Let $\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$, and $R\subseteq C_1\times C_2$. \begin{itemize} \item A distribution $\mu\in\Distr(C_1\times C_2)$ is a coupling for $\mu_1$ and $\mu_2$ iff its first and second marginals coincide with $\mu_1$ and $\mu_2$ respectively, i.e.\, $\Distr(\pi_1)(\mu)=\mu_1$ and $\Distr(\pi_2)(\mu)=\mu_2$. \item A distribution $\mu\in\Distr(C_1\times C_2)$ is a $R$-coupling for $\mu_1$ and $\mu_2$ if it is a coupling for $\mu_1$ and $\mu_2$ and, moreover, $\Pr_{(x_1,x_2)\leftarrow \mu} [R~x_1~x_2]=1$, i.e., if the support of the distribution $\mu$ is included in $R$.\end{itemize} Moreover, we write $\coupl{\mu_1}{\mu_2}{R}$ iff there exists a $R$-coupling for $\mu_1$ and $\mu_2$. \end{definition} Couplings always exist. For instance, the product distribution of two distributions is always a coupling. Going back to the example about the two coins, it can be proven by computation that the following is a coupling that lifts the less-or-equal relation ($0$ indicating heads and $1$ indicating tails): \[\left\{\begin{array}{lll} &(0,0) \ \text{w/ prob}\ (1-q)\quad\quad &(0,1) \ \text{w/ prob}\ (q-p) \\ &(1,0) \ \text{w/ prob}\ 0\quad\quad &(1,1) \ \text{w/ prob}\ p \end{array}\right.\] The following theorem in~\cite{strassen1965existence} gives a necessary and sufficient condition for the existence of $R$-couplings between two distributions. The theorem is remarkable in the sense that it proves an equivalence between an existential property (namely the existence of a particular coupling) and a universal property (checking, for each event, an inequality between probabilities). \begin{theorem}[Strassen's theorem] Consider $\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$, and $R\subseteq C_1 \times C_2$. Then $\coupl{\mu_1}{\mu_2}{R}$ iff for every $X \subseteq C_1$, $\Pr_{x_1\leftarrow \mu_1}[x_1\in X] \leq \Pr_{x_2\leftarrow \mu_2}[x_2\in R(X)]$, where $R(X)$ is the image of $X$ under $R$, i.e.\, $R(X) =\{ y \in C_2 \mid \exists x \in X.~R~x~y\}$. \end{theorem} An important property of couplings is closure under sequential composition. \begin{lemma}[Sequential composition couplings]\label{lem:sequential-composition-of-couplings} Let $\mu_1\in\Distr(C_1)$, $\mu_2\in\Distr(C_2)$, $M_1:C_1\rightarrow \Distr(D_1)$ and $M_2:C_2\rightarrow \Distr(D_2)$. Moreover, let $R\subseteq C_1\times C_2$ and $S\subseteq D_1\times D_2$. Assume: $(1)$ $\coupl{\mu_1}{\mu_2}{R}$; and $(2)$ for every $x_1\in C_1$ and $x_2\in C_2$ such that $R~x_1~x_2$, we have $\coupl{M_1(x_1)}{M_2(x_2)}{S}$. Then $\coupl{(\mathsf{bind}~\mu_1~M_1)}{(\mathsf{bind}~\mu_2~M_2)}{S}$, where $\mathsf{bind}~\mu~M$ is defined as $$(\mathsf{bind}~\mu~M)(y) =\sum_x \mu(x) \cdot M(x)(y)$$ \end{lemma} We conclude this section with the following lemma, which follows from Strassen's theorem: \begin{lemma}[Fundamental lemma of couplings]\label{lem:fun-coup} Let $R\subseteq C_1 \times C_2$, $E_1\subseteq C_1$ and $E_2\subseteq C_2$ such that for every $x_1\in E_1$ and $x_2\in C_2$, $R~x_1~x_2$ implies $x_2\in E_2$, i.e.\, $R(E_1)\subseteq E_2$. Moreover, let $\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$ such that $\coupl{\mu_1}{\mu_2}{R}$. Then $$\Pr_{x_1\leftarrow \mu_1} [x_1\in E_1] \leq \Pr_{x_2\leftarrow\mu_2} [x_2\in E_2]$$ \end{lemma} This lemma can be used to prove probabilistic inequalities from the existence of suitable couplings: \begin{corollary}\label{cor:fundamental} Let $\mu_1,\mu_2\in\Distr(C)$: \begin{enumerate} \item If $\coupl{\mu_1}{\mu_2}{(=)}$, then for all $x\in C$, $\mu_1(x) = \mu_2(x)$. \item If $C = \nat$ and $\coupl{\mu_1}{\mu_2}{(\geq)}$, then for all $n\in \nat$, $\Pr_{x\leftarrow \mu_1}[x\geq n] \geq \Pr_{x\leftarrow \mu_2}[x\geq n]$ \end{enumerate} \end{corollary} In the example at the beginning of the section, the property we want to prove is precisely that, for every $k$ and $i$, the following holds: \[ \Pr_{x_1\leftarrow n_i} [x_1 \geq k] \leq \Pr_{x_2\leftarrow m_i} [x_2 \geq k] \] Since we have a $\leq$-coupling, this proof is immediate. This example is formalized in \autoref{sec:proba-ex}. \section{Overview of the system} \label{sec:overview} In this section we give a high-level overview of our system, with the details on sections \ref{sec:syntax}, \ref{sec:ghol} and \ref{sec:grhol}. We start by presenting the base logic, and then we show how to extend it with probabilities and how to build a relational reasoning system on top of it. \subsection{Base logic: Guarded Higher-Order Logic} Our starting point is the Guarded Higher-Order Logic~\cite{CBGB16} (Guarded HOL) inspired by the topos of trees. In addition to the usual constructs of HOL to reason about lambda terms, this logic features the $\later$ and $\square$ modalities to reason about infinite terms, in particular streams. The $\later$ modality is used to reason about objects that will be available in the future, such as tails of streams. For instance, suppose we want to define an $\All(s,\phi)$ predicate, expressing that all elements of a stream $s \equiv \cons{n}{xs}$ satisfy a property $\phi$. This can be axiomatized as follows: \[\forall (xs: \later \Str{\nat}) (n : \nat). \phi\ n \Rightarrow \later[s\ot xs]{\All(s, x. \phi)} \Rightarrow \All(\cons{n}{xs}, x. \phi)\] We use $x. \phi$ to denote that the formula $\phi$ depends on a free variable $x$, which will get replaced by the first argument of $\All$. We have two antecedents. The first one states that the head $n$ satisfies $\phi$. The second one, $\later[s\ot xs]{\All(s, x. \phi)}$, states that all elements of $xs$ satisfy $\phi$. Formally, $xs$ is the tail of the stream and will be available in the future, so it has type $\later\Str{\nat}$. The \emph{delayed substitution} $\triangleright[s\ot xs]$ replaces $s$ of type $\Str{\nat}$ with $xs$ of type $\later\Str{\nat}$ inside $\All$ and shifts the whole formula one step into the future. In other words, $\later[s\ot xs]{\All(s, x. \phi)}$ states that $\All(-, x.\phi)$ will be satisfied by $xs$ in the future, once it is available. \subsection{A system for relational reasoning} \label{sec:overview:grhol} When proving relational properties it is often convenient to build proofs guided by the syntactic structure of the two expressions to be related. This style of reasoning is particularly appealing when the two expressions have the same structure and control-flow, and is appealingly close to the traditional style of reasoning supported by refinement types. At the same time, a strict adherence to the syntax-directed discipline is detrimental to the expressiveness of the system; for instance, it makes it difficult or even impossible to reason about structurally dissimilar terms. To achieve the best of both worlds, we present a relational proof system built on top of Guarded HOL, which we call Guarded RHOL. Judgements have the shape: \[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi}\] where $\phi$ is a logical formula that may contain two distinguished variables $\res\ltag$ and $\res\rtag$ that respectively represent the expressions $t_1$ and $t_2$. This judgement subsumes two typing judgements on $t_1$ and $t_2$ and a relation $\phi$ on these two expressions. However, this form of judgement does not tie the logical property to the type of the expressions, and is key to achieving flexibility while supporting syntax-directed proofs whenever needed. The proof system combines rules of two different flavours: two-sided rules, which relate expressions with the same top-level constructs, and one-sided rules, which operate on a single expression. We then extend Guarded HOL with a modality $\diamond$ that lifts assertions over discrete types $C_1$ and $C_2$ to assertions over $\Distr(C_1)$ and $\Distr(C_2)$. Concretely, we define for every assertion $\phi$, variables $x_1$ and $x_2$ of type $C_1$ and $C_2$ respectively, and expressions $t_1$ and $t_2$ of type $\Distr(C_1)$ and $\Distr (C_2)$ respectively, the modal assertion $\diamond_{ [x_1\leftarrow t_1,x_2\leftarrow t_2]} \phi$ which holds iff the interpretations of $t_1$ and $t_2$ are related by the probabilistic lifting of the interpretation of $\phi$. We call this new logic Probabilistic Guarded HOL. We accordingly extend the relational proof system to support reasoning about probabilistic expressions by adding judgements of the form: \[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)} {t_2}{\Distr(C_2)}{ \diamond_{[x_1\leftarrow \res\ltag, x_2\leftarrow \res\rtag]} \phi}\] expressing that $t_1$ and $t_2$ are distributions related by a $\phi$-coupling. We call this proof system Probabilistic Guarded RHOL. These judgements can be built by using the following rule, that lifts relational judgements over discrete types $C_1$ and $C_2$ to judgements over distribution types $\Distr(C_1)$ and $\Distr(C_2)$ when the premises of Strassen's theorem are satisfied. \[ \infer[\sf COUPLING]{ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{t_2}{\Distr(C_2)}{ \diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]}\phi}}{ \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall X_1 \subseteq C_1. \Pr_{y_1\ot t_1} [y_1\in X_1] \leq \Pr_{y_2\ot t_2} [\exists y_1 \in X_1. \phi]}} \] Recall that (discrete time) Markov chains are ``memoryless'' probabilistic processes, whose specification is given by a (discrete) set $C$ of states, an initial state $s_0$ and a probabilistic transition function $\mathsf{step}:C \rightarrow \Distr(C)$, where $\Distr(S)$ represents the set of discrete distributions over $C$. As explained in the introduction, a convenient modelling of Markov chains is by means of probabilistic streams, i.e.\, to model a Markov chain as an element of $\Distr(\Str{S})$, where $S$ is its underlying state space. To model Markov chains, we introduce a $\markov$ operator with type $C \to (C \to \Distr(C)) \to \Distr(\Str{C})$ that, given an initial state and a transition function, returns a Markov chain. We can reason about Markov chains by the \rname{Markov} rule (the context, omitted, does not change): \begin{gather*} \inferrule*[right=\sf Markov] {\jgrholnoc{t_1}{C_1}{t_2}{C_2}{\phi} \\\\ \jgrholnoc{h_1}{C_1 \to \Distr(C_1)}{h_2}{C_2 \to \Distr(C_2)}{\psi_3} \\\\ \vdash \psi_4} {\jgrholnoc{\operatorname{markov}(t_1,h_1)}{\Distr(\Str{D_1})}{\operatorname{markov}(t_2,h_2)}{\Distr(\Str{D_2})} {\diamond_{\left[\substack{y_1 \ot \res\ltag \\ y_2 \ot \res\rtag}\right]}\phi'}} \\[0.5em] \text{ where } \begin{cases} \psi_3 \equiv \forall x_1 x_2. \phi\defsubst{x_1}{x_2} \Rightarrow \diamond_{[ y_1 \ot \res\ltag\ x_1, y_2 \ot \res\rtag\ x_2]}\phi\defsubst{y_1}{y_2} \\ \psi_4 \equiv \forall x_1\ x_2\ xs_1\ xs_2. \phi\defsubst{x_1}{x_2} \Rightarrow \later[y_1 \ot xs_1, y_2 \ot xs_2]{\phi'} \Rightarrow \\ \quad\quad\quad\phi'\subst{y_1}{\cons{x_1}{xs_1}}\subst{y_2}{\cons{x_2}{xs_2}} \end{cases} \end{gather*} Informally, the rule stipulates the existence of an invariant $\phi$ over states. The first premise insists that the invariant hold on the initial states, the condition $\psi_3$ states that the transition functions preserve the invariant, and $\psi_4$ states that the invariant $\phi$ over pairs of states can be lifted to a stream property $\phi'$. Other rules of the logic are given in Figure~\ref{fig:phol}. The language construct $\mathsf{munit}$ creates a point distribution whose entire mass is at its argument. Accordingly, the [\textsf{UNIT}] rule creates a straightforward coupling. The [\textsf{MLET}] rule internalizes sequential composition of couplings (Lemma~\ref{lem:sequential-composition-of-couplings}) into the proof system. The construct $\mlet{x}{t}{t'}$ composes a distribution $t$ with a probabilistic computation $t'$ with one free variable $x$ by sampling $x$ from $t$ and running $t'$. The [\textsf{MLET-L}] rule supports one-sided reasoning about $\mlet{x}{t}{t'}$ and relies on the fact that couplings are closed under convex combinations. Note that one premise of the rule uses a unary judgement, with a non-relational modality $\diamond_{[x\leftarrow \res]} \phi$ whose informal meaning is that $\phi$ holds with probability $1$ in the distribution $\res$. The following table summarizes the different base logics we consider,the relational systems we build on top of them, including the ones presented in~\cite{ABGGS17}, and the equivalences between both sides: \begin{center} \small \begin{tabular}{lcl} Relational logic & & Base logic \\ \hline &&\\[-2mm] $\begin{array}{l}\text{RHOL~\cite{ABGGS17}} \\ \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid \phi \end{array} $ & $\stackrel{\text{\cite{ABGGS17}}}{\Longleftrightarrow}$ & $\begin{array}{l} \text{HOL~\cite{ABGGS17}} \\ \Gamma \mid \Psi \vdash \phi\defsubst{t_1}{t_2} \end{array} $ \\[1em] $\begin{array}{l} \text{Guarded RHOL~\S\ref{sec:grhol}} \\ \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid \phi \end{array} $ & $\stackrel{\text{Thm~\ref{thm:equiv-rhol-hol}}}{\Longleftrightarrow}$ & $\begin{array}{l} \text{Guarded HOL~\cite{CBGB16}} \\ \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi\defsubst{t_1}{t_2} \end{array} $ \\[1em] $\begin{array}{l} \text{Probabilistic Guarded RHOL~\S\ref{sec:grhol}} \\ \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid \diamond_{[y_1 \ot \res\ltag, y_2\ot\res\rtag]}.\phi \end{array} $ & \hspace{.3cm} $\stackrel{\text{Thm~\ref{thm:equiv-rhol-hol}}}{\Longleftrightarrow}$ \hspace{.3cm} & $\begin{array}{l} \text{Probabilistic Guarded HOL~\S\ref{sec:ghol}} \\ \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \diamond_{[y_1 \ot t_1, y_2 \ot t_2]}.\phi \end{array} $ \end{tabular} \end{center} \begin{figure*}[!tb] \small \begin{gather*} \infer[\sf UNIT]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\munit{t_1}}{\Distr(C_1)} {\munit{t_2}}{\Distr(C_2)}{ \diamond_{[x_1\leftarrow \res\ltag, x_2\leftarrow \res\rtag]} \phi}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi [\res\ltag/x_1, \res\rtag/x_2]}} \\[0.2em] \infer[\sf MLET]{ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x_1}{t_1}{t'_1}}{ \Distr(D_1)}{\mlet{x_2}{t_2}{t'_2}}{\Distr(D_2)}{\diamond_{\left[\substack{y_1\leftarrow \res_1 \\ y_2\leftarrow \res_2}\right]} \psi}} {\begin{array}{c}\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{t_2}{\Distr(C_2)}{ \diamond_{[x_1\leftarrow \res_1, x_2\leftarrow \res_2]} \phi} \\ \jgrhol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi }{t'_1}{\Distr(D_1)} {t'_2}{\Distr(D_2)}{\diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi} \end{array}} \\[0.2em] \infer[\sf MLET-L]{ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x_1}{t_1}{t'_1}}{\Distr(D_1)}{t'_2}{\Distr(D_2)}{ \diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi}} {\begin{array}{c} \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{\diamond_{[x\leftarrow \res]} \phi} \\ \jgrhol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi }{t'_1}{\Distr(D_1)}{t'_2}{\Distr(D_2)}{\diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi} \end{array}} \end{gather*} \shrinkcaption \caption{Proof rules for probabilistic constructs}\label{fig:phol} \end{figure*} \subsection{Examples}\label{sec:proba-ex} We formalize elementary examples from the literature on security and Markov chains. None of these examples can be verified in prior systems. Uniformity of \emph{one-time pad} and lumping of \emph{random walks} cannot even be stated in prior systems because the two related expressions in these examples have different types. The \emph{random walk vs lazy random walk} (shift coupling) cannot be proved in prior systems because it requires either asynchronous reasoning or code rewriting. Finally, the \emph{biased coin example} (stochastic dominance) cannot be proved in prior work because it requires Strassen's formulation of the existence of coupling (rather than a bijection-based formulation) or code rewriting. We give additional details below. \subsubsection{One-time pad/probabilistic non-interference} Non-interference~\cite{GoguenM82} is a baseline information flow policy that is often used to model confidentiality of computations. In its simplest form, non-interference distinguishes between public (or low) and private (or high) variables and expressions, and requires that the result of a public expression not depend on the value of its private parameters. This definition naturally extends to probabilistic expressions, except that in this case the evaluation of an expression yields a distribution rather than a value. There are deep connections between probabilistic non-interference and several notions of (information-theoretic) security from cryptography. In this paragraph, we illustrate different flavours of security properties for one-time pad encryption. Similar reasoning can be carried out for proving (passive) security of secure multiparty computation algorithms in the 3-party or multi-party setting~\cite{BogdanovNTW12,Cramer:2015:SMC}. One-time pad is a perfectly secure symmetric encryption scheme. Its space of plaintexts, ciphertexts and keys is the set $\{0,1\}^\ell$---fixed-length bitstrings of size $\ell$. The encryption algorithm is parametrized by a key $k$---sampled uniformly over the set of bitstrings $\{ 0,1 \}^\ell$---and maps every plaintext $m$ to the ciphertext $c = k \oplus m$, where the operator $\oplus$ denotes bitwise exclusive-or on bitstrings. We let $\mathsf{otp}$ denote the expression $\lambda m. \mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k\oplus m}}$, where $\mathcal{U}_{X}$ is the uniform distribution over a finite set $X$. One-time pad achieves perfect security, i.e.\, the distributions of ciphertexts is independent of the plaintext. Perfect security can be captured as a probabilistic non-interference property: $$\jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}{ \forall m_1m_2. \res\ltag~m_1 \stackrel{\diamond}{=} \res\rtag~m_2 }$$ where $e_1 \stackrel{\diamond}{=} e_2$ is used as a shorthand for $\diamond_{[y_1\leftarrow e_1, y_2\leftarrow e_2]} y_1 = y_2$. The crux of the proof is to establish $$m_1,m_2: \{ 0,1 \}^\ell \jgrholnoc{\mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{ \mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{ \res\ltag \oplus m_2 \stackrel{\diamond}{=} \res\rtag \oplus m_1 }$$ using the [\textsf{COUPLING}] rule. It suffices to observe that the assertion induces a bijection, so the image of an arbitrary set $X$ under the relation has the same cardinality as $X$, and hence their probabilities w.r.t.\, the uniform distributions are equal. One can then conclude the proof by applying the rules for monadic sequenciation (\rname{MLET}) and abstraction (rule \rname{ABS} in appendix), using algebraic properties of $\oplus$. Interestingly, one can prove a stronger property: rather than proving that the ciphertext is independent of the plaintext, one can prove that the distribution of ciphertexts is uniform. This is captured by the following judgement: $$c_1, c_2: \{ 0,1 \}^\ell \jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)} {\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}{\psi}$$ where $\psi\defeq \forall m_1\,m_2. m_1=m_2\Rightarrow \diamond_{[y_1\leftarrow \res\ltag~m_1, y_2\leftarrow \res\rtag ~m_2]} y_1=c_1 \Leftrightarrow y_2=c_2$. This style of modelling uniformity as a relational property is inspired from~\cite{BartheEGHS17}. The proof is similar to the previous one and omitted. However, it is arguably more natural to model uniformity of the distribution of ciphertexts by the judgement: $$\jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)} {\mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{ \forall m.~ \res\ltag~m \stackrel{\diamond}{=} \res\rtag }$$ This judgement is closer to the simulation-based notion of security that is used pervasively in cryptography, and notably in Universal Composability~\cite{canetti2001universally}. Specifically, the statement captures the fact that the one-time pad algorithm can be simulated without access to the message. It is interesting to note that the judgement above (and more generally simulation-based security) could not be expressed in prior works, since the two expressions of the judgement have different types---note that in this specific case, the right expression is a distribution but in the general case the right expression will also be a function, and its domain will be a projection of the domain of the left expression. The proof proceeds as follows. First, we prove $$ \jgrholnocnot{\mathcal{U}_{\{0,1\}^\ell}}{ \mathcal{U}_{\{0,1\}^\ell}}{ \forall m.~\diamond_{[y_1\leftarrow \res\ltag, y_2\leftarrow \res\rtag]} y_1 \oplus m = y_2}$$ using the [\textsf{COUPLING}] rule. Then, we apply the [\textsf{MLET}] rule to obtain $$\jgrholnocnot {\begin{array}{l}\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\\ \munit{k\oplus m}} \end{array}} {\begin{array}{l}\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\\ \munit{k}}\end{array}} {\diamond_{\left[y_1\leftarrow \res\ltag, y_2\leftarrow \res\rtag \right]} y_1 = y_2}$$ We have $\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k}} \equiv \mathcal{U}_{\{0,1\}^\ell}$; hence by equivalence (rule \rname{Equiv} in appendix), this entails $$ \jgrholnocnot{\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k\oplus m}}} {\mathcal{U}_{\{0,1\}^\ell}}{\diamond_{[y_1\leftarrow \res\ltag, y_2\leftarrow \res\rtag]} y_1 = y_2}$$ We conclude by applying the one-sided rule for abstraction. \subsubsection{Stochastic dominance} Stochastic dominance defines a partial order between random variables whose underlying set is itself a partial order; it has many different applications in statistical biology (e.g.\ in the analysis of the birth-and-death processes), statistical physics (e.g.\ in percolation theory), and economics. First-order stochastic dominance, which we define below, is also an important application of probabilistic couplings. We demonstrate how to use our proof system for proving (first-order) stochastic dominance for a simple Markov process which samples biased coins. While the example is elementary, the proof method extends to more complex examples of stochastic dominance, and illustrates the benefits of Strassen's formulation of the coupling rule over alternative formulations stipulating the existence of bijections (explained later). We start by recalling the definition of (first-order) stochastic dominance for the $\mathbb{N}$-valued case. The definition extends to arbitrary partial orders. \begin{definition}[Stochastic dominance] Let $\mu_1,\mu_2\in \Distr(\mathbb{N})$. We say that $\mu_2$ stochastically dominates $\mu_1$, written $\mu_1\leq_{\mathrm{SD}} \mu_2$, iff for every $n\in\mathbb{N}$, $$\Pr_{x\leftarrow \mu_1}[x\geq n] \leq \Pr_{x\leftarrow \mu_2}[x\geq n]$$ \end{definition} The following result, equivalent to \autoref{cor:fundamental}, characterizes stochastic dominance using probabilistic couplings. \begin{proposition} Let $\mu_1,\mu_2\in \Distr(\mathbb{N})$. Then $\mu_1\leq_{\mathrm{SD}} \mu_2$ iff $\coupl{\mu_1}{\mu_2}{(\leq)}$. \end{proposition} We now turn to the definition of the Markov chain. For $p\in [0,1]$, we consider the parametric $\mathbb{N}$-valued Markov chain $\mathsf{coins} \defeq \markov(0,h)$, with initial state $0$ and (parametric) step function: $$ h \defeq \lambda x. \mlet{b}{\bern{p}}{\munit{x+b}} $$ where, for $p \in [0,1]$, $\bern{p}$ is the Bernoulli distribution on $\{0,1\}$ with probability $p$ for $1$ and $1-p$ for $0$. Our goal is to establish that $\mathsf{coins}$ is monotonic, i.e.\, for every $p_1,p_2\in [0,1]$, $p_1\leq p_2$ implies $\mathsf{coins}~p_1 \leq_{\mathrm{SD}} \mathsf{coins}~p_2$. We formalize this statement as $$ \jgrholnoc{\mathsf{coins}}{[0,1] \rightarrow \Distr(\Str{\mathbb{N}})} {\mathsf{coins}}{[0,1] \rightarrow \Distr(\Str{\mathbb{N}})} {\psi} $$ where $\psi\defeq \forall p_1,p_2. p_1\leq p_2 \Rightarrow \diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]} \All(y_1, y_2, z_1.z_2.z_1\leq z_2)$. The crux of the proof is to establish stochastic dominance for the Bernoulli distribution: $$ p_1:[0,1],p_2:[0,1]\mid p_1\leq p_2 \jgrholnoc{\bern{p_1}} {\Distr(\mathbb{N})} {\bern{p_2}} {\Distr(\mathbb{N})} {\res\ltag\stackrel{\diamond}{\leq} \res\rtag} $$ where we use $e_1 \stackrel{\diamond}{\leq} e_2$ as shorthand for $\diamond_{[y_1 \ot e_1, y_2 \ot e_2]} y_1\leq y_2$. This is proved directly by the \rname{COUPLING} rule and checking by simple calculations that the premise of the rule is valid. We briefly explain how to conclude the proof. Let $h_1$ and $h_2$ be the step functions for $p_1$ and $p_2$ respectively. It is clear from the above that (context omitted): $$ x_1\leq x_2 \jgrholnoc{h_1\ x_1}{\Distr(\bool)} {h_2\ x_2}{\Distr(\bool)} {\diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]}. {y_1\leq y_2}} $$ and by the definition of $\All$: $$x_1 \leq x_2 \Rightarrow {\All(xs_1,xs_2,z_1.z_2.z_1\leq z_2)} \Rightarrow \All(\cons{x_1}{\later xs_1}, \cons{x_2}{\later xs_2}, z_1.z_2.z_1\leq z_2)$$ So, we can conclude by applying the \rname{Markov} rule. It is instructive to compare our proof with prior formalizations, and in particular with the proof in \cite{BartheEGHSS15}. Their proof is carried out in the \textsf{pRHL} logic, whose [\textsf{COUPLING}] rule is based on the existence of a bijection that satisfies some property, rather than on our formalization based on Strassen's Theorem. Their rule is motivated by applications in cryptography, and works well for many examples, but is inconvenient for our example at hand, which involves non-uniform probabilities. Indeed, their proof is based on code rewriting, and is done in two steps. First, they prove equivalence between sampling and returning $x_1$ from $\bern{p_1}$; and sampling $z_1$ from $\bern{p_2}$, $z_2$ from $\bern{\rfrac{p_1}{p_2}}$ and returning $z= z_1 \land z_2$. Then, they find a coupling between $z$ and $\bern{p_2}$. \subsubsection{Shift coupling: random walk vs lazy random walk} The previous example is an instance of a lockstep coupling, in that it relates the $k$-th element of the first chain with the $k$-th element of the second chain. Many examples from the literature follow this lockstep pattern; however, it is not always possible to establish lockstep couplings. Shift couplings are a relaxation of lockstep couplings where we relate elements of the first and second chains without the requirement that their positions coincide. We consider a simple example that motivates the use of shift couplings. Consider the random walk and lazy random walk (which, at each time step, either chooses to move or stay put), both defined as Markov chains over $\mathbb{Z}$. For simplicity, assume that both walks start at position 0. It is not immediate to find a coupling between the two walks, since the two walks necessarily get desynchronized whenever the lazy walk stays put. Instead, the trick is to consider a lazy random walk that moves two steps instead of one. The random walk and the lazy random walk of step 2 are defined by the step functions: $$\begin{array}{rcl} \operatorname{step} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\munit{z+x}} \\ \operatorname{lstep2} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\mlet{b}{\mathcal{U}_{\{0,1\}}}{\munit{x+2*z*b}}} \end{array}$$ After 2 iterations of $\operatorname{step}$, the position has either changed two steps to the left or to the right, or has returned to the initial position, which is the same behaviour $\operatorname{lstep2}$ has on every iteration. Therefore, the coupling we want to find should equate the elements at position $2i$ in $\operatorname{step}$ with the elements at position $i$ in $\operatorname{lstep2}$. The details on how to prove the existence of this coupling are in \autoref{sec:grhol}. \subsubsection{Lumped coupling: random walks on 3 and 4 dimensions} A Markov chain is \emph{recurrent} if it has probability 1 of returning to its initial state, and \emph{transient} otherwise. It is relatively easy to show that the random walk over $\mathbb{Z}$ is recurrent. One can also show that the random walk over $\mathbb{Z}^2$ is recurrent. However, the random walk over $\mathbb{Z}^3$ is transient. For higher dimensions, we can use a coupling argument to prove transience. Specifically, we can define a coupling between a lazy random walk in $n$ dimensions and a random walk in $n +m$ dimensions, and derive transience of the latter from transience of the former. We define the (lazy) random walks below, and sketch the coupling arguments. Specifically, we show here the particular case of the transience of the 4-dimensional random walk from the transience of the 3-dimensional lazy random walk. We start by defining the stepping functions: \[\begin{array}{rl} \operatorname{step}_4 &: \mathbb{Z}^4 \to \Distr(\mathbb{Z}^4) \defeq \lambda z_1. \mlet{x_1}{\unif{U_4}}{\munit{z_1 +_4 x_1}} \\ \operatorname{lstep}_3 &: \mathbb{Z}^3 \to \Distr(\mathbb{Z}^3) \defeq \lambda z_2. \mlet{x_2}{\unif{U_3}}{\mlet{b_2}{\bern{\rfrac{3}{4}}}{\munit{z_2 +_3 b_2*x_2 }}} \end{array} \] where $U_i=\{(\pm 1,0,\dots 0), \dots, (0,\dots,0,\pm 1)\}$ are the vectors of the basis of $\mathbb{Z}^i$ and their opposites. Then, the random walk of dimension 4 is modelled by $\operatorname{rwalk4} \defeq \markov(0, \operatorname{step_4})$, and the lazy walk of dimension 3 is modelled by $\operatorname{lwalk3} \defeq \markov(0, \operatorname{step_3})$. We want to prove: $$ \jgrholnoc{\operatorname{rwalk4}}{\Distr(\Str{\mathbb{Z}^{4}})} {\operatorname{lwalk3}}{\Distr(\Str{\mathbb{Z}^{3}})} {\diamond_{\left[\substack{y_1\leftarrow \res\ltag \\ y_2\leftarrow \res\rtag}\right]} \All(y_1, y_2, z_1.z_2.\operatorname{pr}^{4}_{3}(z_1) = z_2)} $$ where $\operatorname{pr}^{n_2}_{n_1}$ denotes the standard projection from $\mathbb{Z}^{n_2}$ to $\mathbb{Z}^{n_1}$. We apply the \rname{Markov} rule. The only interesting premise requires proving that the transition function preserves the coupling: \[p_2=\operatorname{pr}^{4}_{3}(p_1) \vdash \operatorname{step_4} \sim \operatorname{lstep}_3 \mid \forall x_1 x_2. x_2=\operatorname{pr}^4_3(x_1) \Rightarrow \diamond_{\left[\substack{y_1\ot \res\ltag\ x_1 \\ y_2\ot \res\rtag\ x_2}\right]} \operatorname{pr}^4_3(y_1)=y_2 \] To prove this, we need to find the appropriate coupling, i.e., one that preserves the equality. The idea is that the step in $\mathbb{Z}^3$ must be the projection of the step in $\mathbb{Z}^4$. This corresponds to the following judgement: \[\left.\begin{array}{rl} \lambda z_1. &\mlet{x_1}{\unif{U_4}} {\\ &\munit{z_1 +_4 x_1}} \end{array} \sim \begin{array}{rl} \lambda z_2. &\mlet{x_2}{\unif{U_3}} {\\ &\mlet{b_2}{\bern{\rfrac{3}{4}}} {\\ &\munit{z_2 +_3 b_2*x_2}}} \end{array} \; \right| \; \begin{array}{c} \forall z_1 z_2. \operatorname{pr}^4_3(z_1) = z_2 \Rightarrow \\ \operatorname{pr}^4_3(\res\ltag\ z_1) \stackrel{\diamond}{=} \res\rtag\ z_2 \end{array}\] which by simple equational reasoning is the same as \[\left.\begin{array}{rl} \lambda z_1. &\mlet{x_1}{\unif{U_4}} {\\ &\munit{z_1 +_4 x_1}} \end{array} \sim \begin{array}{rl} \lambda z_2. &\mlet{p_2}{\unif{U_3} \times \bern{\rfrac{3}{4}}} {\\ &\munit{z_2 +_3 \pi_1(p_2)*\pi_2(p_2)}} \end{array} \; \right| \; \begin{array}{c} \forall z_1 z_2. \operatorname{pr}^4_3(z_1) = z_2 \Rightarrow \\ \operatorname{pr}^4_3(\res\ltag\ z_1) \stackrel{\diamond}{=} \res\rtag\ z_2 \end{array} \] We want to build a coupling such that if we sample $(0,0,0,1)$ or $(0,0,0,-1)$ from $\unif{U_3}$, then we sample $0$ from $\bern{\rfrac{3}{4}}$, and otherwise if we sample $(x_1,x_2,x_3,0)$ from $\unif{U_4}$, we sample $(x_1,x_2,x_3)$ from $U_3$. Formally, we prove this with the \rname{Coupling} rule. Given $X:U_4 \to \bool$, by simple computation we show that: \[ \Pr_{z_1 \sim \unif{U_4}}[z_1 \in X] \leq \Pr_{z_2 \sim \unif{U_3} \times \bern{\rfrac{3}{4}}}[z_2 \in \{ y \mid \exists x \in X. {\sf pr}^4_3(x) = \pi_1(y)*\pi_2(y) \}]\] This concludes the proof. From the previous example, it follows that the lazy walk in 3 dimensions is transient, since the random walk in 3 dimensions is transient. By simple reasoning, we now conclude that the random walk in 4 dimensions is also transient. \section{Probabilistic Guarded Lambda Calculus} \label{sec:syntax} To ensure that a function on infinite datatypes is well-defined, one must check that it is \emph{productive}. This means that any finite prefix of the output can be computed in finite time. For instance, consider the following function on streams: \[ \mathtt{letrec\ bad\ (x:xs) = x : tail (bad\ xs)} \] This function is not productive since only the first element can be computed. We can argue this as follows: Suppose that the tail of a stream is available one unit of time after its head, and that that \texttt{x:xs} is available at time 0. How much time does it take for \texttt{bad} to start outputting its tail? Assume it takes $k$ units of time. This means that \texttt{tail(bad\ xs)} will be available at time $k+1$ , since \texttt{xs} is only available at time 1. But \texttt{tail(bad\ xs)} is exactly the tail of \texttt{bad(x:xs)}, and this is a contradiction, since \texttt{x:xs} is available at time 0 and therefore the tail of \texttt{bad(x:xs)} should be available at time $k$. Therefore, the tail of \texttt{bad} will never be available. The guarded lambda calculus solves the productivity problem by distinguishing at type level between data that is available now and data that will be available in the future, and restricting when fixpoints can be defined. Specifically, the guarded lambda calculus extends the usual simply typed lambda calculus with two modalities: $\later$ (pronounced \textit{later}) and $\square$ (\textit{constant}). The later modality represents data that will be available one step in the future, and is introduced and removed by the term formers $\later$ and $\prev$\! respectively. This modality is used to guard recursive occurrences, so for the calculus to remain productive, we must restrict when it can be eliminated. This is achieved via the constant modality, which expresses that all the data is available at all times. In the remainder of this section we present a probabilistic extension of this calculus. \paragraph*{Syntax} Types of the calculus are defined by the grammar \begin{align*} A,B ::= b \mid \nat \mid A \times B \mid A + B \mid A \to B \mid \Str{A} \mid \square~A \mid \later{A} \mid \Distr(C) \end{align*} where $b$ ranges over a collection of base types. $\Str{A}$ is the type of guarded streams of elements of type $A$. Formally, the type $\Str{A}$ is isomorphic to $A \times \later{\Str{A}}$. This isomorphism gives a way to introduce streams with the function $(\cons{}{}) : A \to \later\Str{A} \to \Str{A}$ and to eliminate them with the functions $\hd: \Str{A} \to A$ and $\tl: \Str{A} \to \later\Str{A}$. $\Distr (C)$ is the type of distributions over \emph{discrete types} $C$. Discrete types are defined by the following grammar, where $b_0$ are discrete base types, e.g., $\zint$. \begin{align*} C,D ::= b_0 \mid \nat \mid C \times D \mid C + D \mid \Str{C} | \later{C}. \end{align*} Note that, in particular, arrow types are not discrete but streams are. This is due to the semantics of streams as sets of finite approximations, which we describe in the next subsection. Also note that $\square\Str{A}$ is not discrete since it makes the full infinite streams available. We also need to distinguish between arbitrary types $A, B$ and constant types $S, T$, which are defined by the following grammar \begin{align*} S, T ::= b_C \mid \nat \mid S \times T \mid S + T \mid S \to T \mid \square~A \end{align*} where $b_C$ is a collection of constant base types. Note in particular that for any type $A$ the type $\square~A$ is constant. The terms of the language $t$ are defined by the following grammar \begin{align*} t &::= ~x \mid c \mid 0 \mid S t \mid \casenat{t}{t}{t} \mid \mu \mid \munit{t} \mid \mlet{x}{t}{t} \\ &\mid \langle t, t \rangle \mid \pi_1 t \mid \pi_2 t \mid \inj{1} t \mid \inj{2} t \mid \CASE t \OF \inj{1} x . t; \inj{2} y . t \mid \lambda x . t \mid t\,t \mid \fix{x}{t}\\ &\mid \cons{t}{ts} \mid \hd t \mid \tl t \mid \boxx{t} \mid \letbox{x}{t}{t} \mid \letconst{x}{t}{t} \mid \latern[\xi]t \mid \prev t \end{align*} where $\xi$ is a delayed substitution, a sequence of bindings $\hrt{x_1 \gets t_1, \ldots, x_n \gets t_n}$. The terms $c$ are constants corresponding to the base types used and $\munit{t}$ and $\mlet{x}{t}{t}$ are the introduction and sequencing construct for probability distributions. The meta-variable $\mu$ stands for base distributions like $\unif{C}$ and $\bern{p}$. Delayed substitutions were introduced in \cite{BGCMB16} in a dependent type theory to be able to work with types dependent on terms of type $\later{A}$. In the setting of a simple type theory, such as the one considered in this paper, delayed substitutions are equivalent to having the applicative structure~\cite{McBride:Applicative} $\app$ for the $\later$ modality. However, delayed substitutions extend uniformly to the level of propositions, and thus we choose to use them in this paper in place of the applicative structure. \paragraph*{Denotational semantics} The meaning of terms is given by a denotational model in the category $\trees$ of presheaves over $\omega$, the first infinite ordinal. This category $\trees$ is also known as the \emph{topos of trees}~\cite{Birkedal-et-al:topos-of-trees}. In previous work~\cite{CBGB16}, it was shown how to model most of the constructions of the guarded lambda calculus and its internal logic, with the notable exception of the probabilistic features. Below we give an elementary presentation of the semantics. Informally, the idea behind the topos of trees is to represent (infinite) objects from their finite approximations, which we observe incrementally as time passes. Given an object $x$, we can consider a sequence $\{x_i\}$ of its finite approximations observable at time $i$. These are trivial for finite objects, such as a natural number, since for any number $n$, $n_i = n$ at every $i$. But for infinite objects such as streams, the $i$th approximation is the prefix of length $i+1$. Concretely, the category $\trees$ consists of: \begin{itemize} \item Objects $X$: families of sets $\{X_i\}_{i\in\nat}$ together with \emph{restriction functions} $r_n^X : X_{n+1} \to X_n$. We will write simply $r_n$ if $X$ is clear from the context. \item Morphisms $X \to Y$ : families of functions $\alpha_n : X_n \to Y_n$ commuting with restriction functions in the sense of $r_n^Y \circ \alpha_{n+1} = \alpha_n \circ r_n^X$. \end{itemize} The full interpretation of types of the calculus can be found in \autoref{fig:sem-types} in the appendix. The main points we want to highlight are: \begin{itemize} \item Streams over a type $A$ are interpreted as sequences of finite prefixes of elements of $A$ with the restriction functions of $A$: $$\sem{\Str{A}} \defeq \sem{A}_0 \times \{*\} \xleftarrow{r_0 \times !} \sem{A}_1 \times \sem{\Str{A}}_0 \xleftarrow{r_1 \times r_0 \times !} \sem{A}_2\times \sem{\Str{A}}_1 \leftarrow \cdots$$ \item Distributions over a discrete object $C$ are defined as a sequence of distributions over each $\sem{C}_i$: $$\sem{\Distr(C)} \defeq \Distr(\sem{C}_0) \stackrel{\Distr(r_0)}{\longleftarrow} \Distr(\sem{C}_1) \stackrel{\Distr(r_1)}{\longleftarrow} \Distr(\sem{C}_2) \stackrel{\Distr(r_2)}{\longleftarrow} \ldots,$$ where $\Distr(\sem{C}_i)$ is the set of (probability density) functions $\mu : \sem{C}_i \to [0,1]$ such that $\sum_{x_\in X} \mu x = 1$, and $\Distr(r_i)$ adds the probability density of all the points in $\sem{C}_{i+1}$ that are sent by $r_i$ to the same point in the $\sem{C}_{i}$. In other words, $\Distr(r_i)(\mu)(x) = \Pr_{y \ot \mu}[r_i(y) = x]$ \end{itemize} An important property of the interpretation is that discrete types are interpreted as objects $X$ such that $X_i$ is finite or countably infinite for every $i$. This allows us to define distributions on these objects without the need for measure theory. In particular, the type of guarded streams $\Str{A}$ is discrete provided $A$ is, which is clear from the interpretation of the type $\Str{A}$. Conceptually this holds because $\sem{\Str{A}}_i$ is an approximation of real streams, consisting of only the first $i+1$ elements. An object $X$ of $\trees$ is \emph{constant} if all its restriction functions are bijections. Constant types are interpreted as constant objects of $\trees$ and for a constant type $A$ the objects $\sem{\square A}$ and $\sem{A}$ are isomorphic in $\trees$. \paragraph*{Typing rules} Terms are typed under a dual context $\Delta \mid \Gamma$, where $\Gamma$ is a usual context that binds variables to a type, and $\Delta$ is a constant context containing variables bound to types that are \textit{constant}. The term $\letconst{x}{u}{t}$ allows us to shift variables between constant and non-constant contexts. The typing rules can be found in \autoref{fig:glambda}. The semantics of such a dual context $\Delta \mid \Gamma$ is given as the product of types in $\Delta$ and $\Gamma$, except that we implicitly add $\square$ in front of every type in $\Delta$. In the particular case when both contexts are empty, the semantics of the dual context correspond to the terminal object $1$, which is the singleton set $\{\ast\}$ at each time. The interpretation of the well-typed term $\Delta \mid \Gamma \vdash t : A$ is defined by induction on the typing derivation, and can be found in \autoref{fig:sem-glc} in the appendix. \begin{figure*}[!tb] \small \begin{gather*} \inferrule {x : A \in \Gamma} {\Delta \mid \Gamma \vdash x : A} \qquad \inferrule {x : A \in \Delta} {\Delta \mid \Gamma \vdash x : A} \qquad \inferrule {\Delta \mid \Gamma, x : A \vdash t : B} {\Delta \mid \Gamma \vdash \lambda x . t : A \to B} \\ \inferrule {\Delta \mid \Gamma \vdash t : A \to B \and \Delta \mid \Gamma \vdash u : A} {\Delta \mid \Gamma \vdash t\,u : B} \qquad \inferrule {\Delta \mid \Gamma, f : \later{A} \vdash t : A} {\Delta \mid \Gamma \vdash \fix{f}{t} : A} \qquad \inferrule {\Delta \mid \cdot \vdash t : \later A} {\Delta \mid \Gamma \vdash \prev{t} : A} \\ \inferrule {\Delta \mid \cdot \vdash t : A} {\Delta \mid \Gamma \vdash \boxx{t} : \square A} \qquad \inferrule {\Delta \mid \Gamma \vdash u : \square B \and \Delta, x : B \mid \Gamma \vdash t : A} {\Delta \mid \Gamma \vdash \letbox{x}{u}{t} : A} \\ \inferrule {\Delta \mid \Gamma \vdash u : B \and \Delta, x : B \mid \Gamma \vdash t : A \and B \text{ constant}} {\Delta \mid \Gamma \vdash \letconst{x}{u}{t} : A} \\ \inferrule {\Delta \mid \Gamma, x_1 : A_1, \cdots x_n : A_n \vdash t : A \and \Delta \mid \Gamma \vdash t_i : \later{A_i}} {\Delta \mid \Gamma \vdash \later[x_1 \gets t_1, \ldots, x_n \gets t_n]{t} : \later{A}} \qquad \inferrule {\Delta \mid \Gamma \vdash t : A \and A \text{ discrete }} {\Delta \mid \Gamma \vdash \munit{t} : \Distr(A)} \\ \inferrule {\Delta \mid \Gamma \vdash t : \Distr(A) \and \Delta \mid \Gamma, x : A \vdash u : \Distr(B)} {\Delta \mid \Gamma \vdash \mlet{x}{t}{u} : \Distr(B)} \qquad \inferrule {\mu \text{ primitive distribution on type } A} {\Delta \mid \Gamma \vdash \mu : \Distr(A)} \end{gather*} \shrinkcaption \caption{A selection of the typing rules of the guarded lambda calculus. The rules for products, sums, and natural numbers are standard.} \label{fig:glambda} \end{figure*} \paragraph*{Applicative structure of the later modality} As in previous work we can define the operator $\app$ satisfying the typing rule \[ \inferrule {\Delta \mid \Gamma \vdash t : \later{(A \to B)} \and \Delta \mid \Gamma \vdash u : \later{A}} {\Delta \mid \Gamma \vdash t \app u : \later{B}} \] and the equation $(\later{t}) \app (\later{u}) \equiv \later(t\ u)$ as the term $t \app u \defeq \latern[\hrt{f \gets t, x \gets u}]{f\,x}$. \paragraph*{Example: Modelling Markov chains} As an application of $\app$ and an example of how to use guardedness and probabilities together, we now give the precise definition of the $\markov$ construct that we used to model Markov chains earlier: \[ \begin{array}{rl} \markov &: C \to (C \to \Distr(C)) \to \Distr(\Str{C}) \\ \markov &\defeq \fix{f}{\lambda x. \lambda h. \\ &\quad \mlet {z}{h\ x}{ \mlet{t}{\operatorname{swap_{\later\Distr}^{\Str{C}}}(f \app \later z \app \later h)} { \munit{\cons{x}{t}}}}} \end{array} \] The guardedness condition gives $f$ the type $\later( C \to (C \to \Distr(C)) \to \Distr(\Str{C}))$ in the body of the fixpoint. Therefore, it needs to be applied functorially (via $\app$) to $\later z$ and $\later h$, which gives us a term of type $\later \Distr(\Str{C})$. To complete the definition we need to build a term of type $\Distr(\later\Str{C})$ and then sequence it with $\cons{}{}$ to build a term of type $\Distr(\Str{C})$. To achieve this, we use the primitive operator $\operatorname{swap_{\later\Distr}^C} : \later\Distr(C) \to \Distr(\later C)$, which witnesses the isomorphism between $\later\Distr(C)$ and $\Distr(\later C)$. For this isomorphism to exist, it is crucial that distributions be total (i.e., we cannot use subdistributions). Indeed, the denotation for $\later\Distr(C)$ is the sequence $\{\ast\} \ot \Distr(C_1) \ot \Distr(C_2) \ot \dots$, while the denotation for $\Distr(\later C)$ is the sequence $\Distr(\{\ast\}) \ot \Distr(C_1) \ot \Distr(C_2) \ot \dots$, and $\{\ast\}$ is isomorphic to $\Distr(\{\ast\})$ in ${\sf Set}$ only if $\Distr$ considers only total distributions. \section{Guarded higher-order logic} \label{sec:ghol} We now introduce Guarded HOL (GHOL), which is a higher-order logic to reason about terms of the guarded lambda calculus. The logic is essentially that of~\cite{CBGB16}, but presented with the dual context formulation analogous to the dual-context typing judgement of the guarded lambda calculus. Compared to standard intuitionistic higher-order logic, the logic GHOL has two additional constructs, corresponding to additional constructs in the guarded lambda calculus. These are the later modality ($\later$) \emph{on propositions}, with delayed substitutions, which expresses that a proposition holds one time unit into the future, and the ``always'' modality $\square$, which expresses that a proposition holds at all times. Formulas are defined by the grammar: \[ \phi, \psi ::= \top \mid \phi \wedge \psi \mid \phi \vee \psi \mid \neg \psi \mid \forall x. \phi \mid \exists x. \phi \mid \later[x_1 \ot t_1 \dots x_n \ot t_n]\phi \mid \square \phi\] The basic judgement of the logic is $\Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi$ where $\Sigma$ is a logical context for $\Delta$ (that is, a list of formulas well-formed in $\Delta$) and $\Psi$ is another logical context for the dual context $\Delta\mid\Gamma$. The formulas in context $\Sigma$ must be \emph{constant} propositions. We say that a proposition $\phi$ is \emph{constant} if it is well-typed in context $\Delta \mid \cdot$ and moreover if every occurrence of the later modality in $\phi$ is under the $\square$ modality. Selected rules are displayed in Figure~\ref{fig:hol} on page~\pageref{fig:hol}. We highlight \rname{Loeb} induction, which is the key to reasoning about fixpoints: to prove that $\phi$ holds now, one can assume that it holds in the future. The interpretation of the formula $\Delta \mid \Gamma \vdash \phi$ is a subobject of the interpretation $\sem{\Delta \mid \Gamma}$. Concretely the interpretation $A$ of $\Delta \mid \Gamma \vdash \phi$ is a family $\left\{A_i\right\}_{i=0}^\infty$ of sets such that $A_i \subseteq \sem{\Delta \mid \Gamma}_i$. This family must satisfy the property that if $x \in A_{i+1}$ then $r_i(x) \in A_i$ where $r_i$ are the restriction functions of $\sem{\Delta \mid \Gamma}$. The interpretation of formulas is defined by induction on the typing derivation. In the interpretation of the context $\Delta \mid \Sigma \mid \Gamma \mid \Psi$ the formulas in $\Sigma$ are interpreted with the added $\square$ modality. Moreover all formulas $\phi$ in $\Sigma$ are typeable in the context $\Delta \mid \cdot \vdash \phi$ and thus their interpretations are subsets of $\sem{\square \Delta}$. We treat these subsets of $\sem{\Delta \mid \Gamma}$ in the obvious way. The cases for the semantics of the judgement $\Delta \mid \Gamma \vdash \phi$ can be found in the appendix. It can be shown that this logic is sound with respect to its model in the topos of trees. \begin{theorem}[Soundness of the semantics]\label{thm:sound-ghol} The semantics of guarded higher-order logic is sound: if $\Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi$ is derivable then for all $n \in \nat$, $\sem{\square\Sigma}_n \cap \sem{\Psi}_n \subseteq \sem{\phi}$. \end{theorem} In addition, Guarded HOL is expressive enough to axiomatize standard probabilities over discrete sets. This axiomatization can be used to define the $\diamond$ modality directly in Guarded HOL (as opposed to our relational proof system, were we use it as a primitive). Furthermore, we can derive from this axiomatization additional rules to reason about couplings, which can be seen in Figure~\ref{fig:prob-hol}. These rules will be the key to proving the soundness of the probabilistic fragment of the relational proof system, and can be shown to be sound themselves. \begin{proposition}[Soundness of derived rules]\label{thm:sound-prob-ghol} The additional rules are sound. \end{proposition} \begin{figure*}[!tb] \small \begin{gather*} \infer[\sf AX_U] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}} {\phi \in \Psi} \quad \infer[\sf AX_G] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}} {\phi \in \Sigma} \quad \infer[\sf CONV] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{t=t'}} {\Gamma \vdash t:\tau & \Gamma \vdash t':\tau & t \equiv t'} \\[0.3em] \infer[\sf SUBST] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x}{u}}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x}{t}} & \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{t = u}} \quad\quad \infer[\sf Loeb] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi, \later\phi}{\phi}} \\[0.3em] \infer[\sf \later_I] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot t_1,\dots,x_n \ot t_n]\phi}} {\jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi}{\phi} & \Delta\mid\Gamma\vdash t_1 : \later A_1 & \dots & \Delta\mid\Gamma\vdash t_n : \later A_n} \\[0.3em] \infer[\sf \later_E] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x_1}{\prev t_1}\dots\subst{x_n}{\prev t_n}}} {\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\later[x_1 \ot t_1 \dots x_n \ot t_n]\phi} & \Delta \mid \bullet \vdash t_1 : \later A_1 & \dots & \Delta \mid \bullet \vdash t_n : \later A_n} \\[0.3em] \inferrule {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1\!\ot\!t_1,\dots,x_n\!\ot\!t_n]\psi} \\ \Delta \mid \Gamma \vdash t_1 : \later A_1 \; \dots\; \Delta \mid \Gamma \vdash t_n : \later A_n \\\\ \jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi, \psi}{\phi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot t_1,\dots,x_n \ot t_n]\phi}} {\sf \later_{App}} \\[0.3em] \infer[\sf \square_I] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\phi}} {\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\phi}} \quad\quad \infer[\sf \square_E] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\psi} & \jghol{\Delta}{\Sigma,\psi}{\Gamma}{\Psi}{\phi}} \end{gather*} \shrinkcaption \caption{Selected Guarded Higher-Order Logic rules}\label{fig:hol} \end{figure*} \begin{figure*}[!tb] \small \begin{gather*} \inferrule*[right = \sf MONO2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2 \leftarrow t_2]}\phi} \\ \jghol{\Delta}{\Sigma}{\Gamma, x_1 : C_1, x_2 : C_2}{\Psi, \phi}{\psi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]}\psi}} \\[0.3em] \inferrule*[right = \sf UNIT2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x_1}{t_1}\subst{x_2}{t_2}}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi} {\diamond_{[x_1\leftarrow \munit{t_1}, x_2\leftarrow \munit{t_2}]}\phi}} \\[0.3em] \inferrule*[right = \sf MLET2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]} \phi} \\ \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi} {\diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi} {\diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow \mlet{x_2}{t_2}{t'_2}]} \psi}} \\[0.3em] \inferrule*[right = \sf MLET{-}L] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1]}\phi} \\ \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi} {\diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi} {\diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow t'_2]} \psi}} \end{gather*} \shrinkcaption \caption{Derived rules for probabilistic constructs}\label{fig:prob-hol} \end{figure*} \section{Relational proof system} \label{sec:grhol} We complete the formal description of the system by describing the proof rules for the non-probabilistic fragment of the relational proof system (the rules of the probabilistic fragment were described in Section~\ref{sec:overview:grhol}). \subsection{Proof rules} The rules for core $\lambda$-calculus constructs are identical to those of~\cite{ABGGS17}; for convenience, we present a selection of the main rules in Figure \ref{fig:sel-rhol} in the appendix. We briefly comment on the two-sided rules for the new constructs (Figure~\ref{fig:s-rhol}). The notation $\Omega$ abbreviates a context $\Delta\mid\Sigma\mid\Gamma\mid\Psi$. The rule \rname{Next} relates two terms that have a $\later$ term constructor at the top level. We require that both have one term in the delayed substitutions and that they are related pairwise. Then this relation is used to prove another relation between the main terms. This rule can be generalized to terms with more than one term in the delayed substitution. The rule \rname{Prev} proves a relation between terms from the same delayed relation by applying $\mathrm{prev}$ to both terms. The rule \rname{Box} proves a relation between two boxed terms if the same relation can be proven in a constant context. Dually, \rname{LetBox} uses a relation between two boxed terms to prove a relation between their unboxings. \rname{LetConst} is similar to \rname{LetBox}, but it requires instead a relation between two constant terms, rather than explicitly $\square$-ed terms. The rule \rname{Fix} relates two fixpoints following the \rname{Loeb} rule from Guarded HOL. Notice that in the premise, the fixpoints need to appear in the delayed substitution so that the inductive hypothesis is well-formed. The rule \rname{Cons} proves relations on streams from relations between their heads and tails, while \rname{Head} and \rname{Tail} behave as converses of \rname{Cons}. Figure~\ref{fig:a-rhol} contains the one-sided versions of the rules. We only present the left-sided versions as the right-sided versions are completely symmetric. The rule \rname{Next-L} relates at $\phi$ a term that has a $\later$ with a term that does not have a $\later$. First, a unary property $\phi'$ is proven on the term $u$ in the delayed substitution, and it is then used as a premise to prove $\phi$ on the terms with delays removed. Rules for proving unary judgements can be found in the appendix. Similarly, \rname{LetBox-L} proves a unary property on the term that gets unboxed and then uses it as a precondition. The rule \rname{Fix-L} builds a fixpoint just on the left, and relates it with an arbitrary term $t_2$ at a property $\phi$. Since $\phi$ may contain the variable $\res\rtag$ which is not in the context, it has to be replaced when adding $\later\phi$ to the logical context in the premise of the rule. The remaining rules are similar to their two-sided counterparts. \begin{figure*}[!htb] \small \begin{gather*} \mbox{}\hspace{-5mm}\inferrule*[right=\sf Next] {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_2}{\Psi, \phi'\defsubst{x_1}{x_2}}{t_1}{A_1}{t_2}{A_2}{\phi} \\ \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{A_1}}{u_2}{\later{A_2}} {\dsubst{\res\ltag, \res\rtag}{\res\ltag, \res\rtag}{\phi'}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi} {\nextt{x_1\!\leftarrow\!u_1}{t_1}}{\later{A_1}} {\nextt{x_2\!\leftarrow\!u_2}{t_2}}{\later{A_2}} {\triangleright[x_1\!\leftarrow\!u_1,x_2\!\leftarrow\!u_2, \res\ltag\!\leftarrow\!\res\ltag,\res\rtag\!\leftarrow\!\res\rtag].\phi}} \\[0.3em] \inferrule*[right=\sf Prev] {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{\later A_2}{\dsubst{\res\ltag,\res\rtag}{\res\ltag,\res\rtag}{\phi}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t_1}}{A_1}{\prev{t_2}}{A_2}{\phi}} \\[0.3em] \mbox{}\hspace{-2mm}\inferrule*[Right=\sf Box] {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{A_1}{t_2}{A_2}{\phi}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t_1}}{\square A_1}{\boxx{t_2}}{\square A_2} {\square \phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}}} \\[0.3em] \inferrule*[right=\sf LetBox] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{u_2}{\square B_2}{\square \phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}} \\ \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{\letbox{x_2}{u_2}{t_2}}{A_2}{\phi'}} \\[0.3em] \inferrule*[right=\sf LetConst] {B_1,B_2,\phi\ \text{constant} \\ FV(\phi)\cap FV(\Gamma) = \emptyset \\ \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{u_2}{B_2}{\phi} \\ \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{\letconst{x_2}{u_2}{t_2}}{A_2}{\phi'}} \\[0.3em] \inferrule*[right=\sf Fix] {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi, \dsubst{\res\ltag,\res\rtag}{f_1,f_2}{\phi}} {t_1}{A_1}{t_2}{A_2}{\phi}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\fix{f_1}{t_1}}{A_1}{\fix{f_2}{t_2}}{A_2}{\phi}} \\[0.3em] \mbox{}\hspace{-5.2mm}\inferrule*[Right=\sf Cons] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{x_2}{A_2}{\phi_h} \\ \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{xs_2}{\later{\Str{A_2}}}{\phi_t} \\ \jgholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall x_1,x_2, s_1, s_2. \phi_h\defsubst{x_1}{x_2} \Rightarrow \phi_t\defsubst{s_1}{s_2} \Rightarrow \phi\defsubst{\cons{x_1}{s_1}}{\cons{x_2}{s_2}}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{s_1}}{\Str{A_1}}{\cons{x_2}{s_2}}{\Str{A_2}}{\phi}} \\[0.3em] \inferrule*[right=\sf Head] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{\Str{A_1}}{\phi\defsubst{hd\ \res\ltag}{hd\ \res\rtag}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{hd\ t_2}{A_2}{\phi}} \\[0.3em] \inferrule*[right=\sf Tail] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{\Str{A_2}}{\phi\defsubst{tl\ \res\ltag}{tl\ \res\rtag}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{\Str{A_1}}}{tl\ t_2}{\later{\Str{A_2}}}{\phi}} \end{gather*} \shrinkcaption \caption{Two-sided rules for Guarded RHOL}\label{fig:s-rhol} \end{figure*} \begin{figure*}[!htb] \small \begin{gather*} \inferrule*[right=\sf Next{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:B_1}{\Psi, \phi'\subst{\res}{x_1}}{t_1}{A_1}{t_2}{A_2}{\phi} \\ \jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{B_1}}{\dsubst{\res}{\res}{\phi'}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi} {\later[x_1 \ot u_1]{t_1}}{\later{A_1}} {t_2}{A_2} {\later[x_1 \ot u_1,\res\ltag\leftarrow\res\ltag]{\phi}}} \\[0.3em] \inferrule*[right=\sf Prev{-}L ] {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{A_2}{\dsubst{\res\ltag}{\res\ltag}{\phi}}} {\jgrhol{\Delta}{\Sigma}{\Gamma_1; \Gamma_2}{\Psi_1 ; \Psi_2}{\prev{t_1}}{A_1}{t_2}{A_2}{\phi}} \\[0.3em] \inferrule*[right=\sf Box{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma_2}{\Psi_2}{t_1}{A_1}{t_2}{A_2}{\phi} \\ FV(t_1)\not\subseteq FV(\Gamma_2)\\ FV(\Psi_2)\subseteq FV(\Gamma_2)} {\jgrhol{\Delta}{\Sigma}{\Gamma_1; \Gamma_2}{\Psi_1; \Psi_2}{\boxx{t_1}}{\square A_1}{t_2}{A_2} {\square \phi\subst{\res\ltag}{\letbox{x_1}{\res\ltag}{x_1}}}} \\[0.3em] \inferrule*[right=\sf LetBox{-}L] {\jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{\square \phi\subst{\res}{\letbox{x_1}{\res\ltag}{x_1}}} \\ \jgrhol{\Delta, x_1 : B_1}{\Sigma, \phi\subst{\res}{x_1}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{t_2}{A_2}{\phi'}} \\[0.3em] \inferrule*[right=\sf LetConst{-}L] {B_1,\phi\ \text{constant} \\ FV(\phi)\cap FV(\Gamma) = \emptyset \\ \jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{\phi} \\\\ \jgrhol{\Delta, x_1 : B_1}{\Sigma, \phi\subst{\res}{x_1}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{t_2}{A_2}{\phi'}} \\[0.3em] \inferrule*[right=\sf Fix{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}}{\Psi, \dsubst{\res\ltag}{f_1}{(\phi\subst{\res\rtag}{t_2})}} {t_1}{A_1}{t_2}{A_2}{\phi}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\fix{f_1}{t_1}}{A_1}{t_2}{A_2}{\phi}} \\[0.3em] \inferrule*[right=\sf Cons{-}L] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{t_2}{A_2}{\phi_h} \\ \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{t_2}{A_2}{\phi_t} \\ \jgholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall x_1,x_2, xs_1. \phi_h\defsubst{x_1}{x_2} \Rightarrow \phi_t\defsubst{xs_1}{x_2} \Rightarrow \phi\defsubst{\cons{x_1}{xs_1}}{x_2}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{xs_1}}{\Str{A_1}}{t_2}{A_2}{\phi}} \\[0.3em] \inferrule*[right=\sf Head{-}L] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{A_2}{\phi\subst{\res\ltag}{hd\ \res\ltag}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{t_2}{A_2}{\phi}} \\[0.3em] \inferrule*[right=\sf Tail{-}L] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{A_2}{\phi\subst{\res\ltag}{tl\ \res\ltag}}} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{\Str{A_1}}}{t_2}{A_2}{\phi}} \end{gather*} \shrinkcaption \caption{One-sided rules for Guarded RHOL}\label{fig:a-rhol} \end{figure*} \subsection{Metatheory} We review some of the most interesting metatheoretical properties of our relational proof system, highlighting the equivalence with Guarded HOL. \begin{theorem}[Equivalence with Guarded HOL] \label{thm:equiv-rhol-hol} For all contexts $\Delta,\Gamma$; types $\sigma_1,\sigma_2$; terms $t_1,t_2$; sets of assertions $\Sigma,\Psi,$; and assertions $\phi$: \[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi} \quad\Longleftrightarrow\quad \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res\ltag}{t_1}\subst{\res\rtag}{t_2}}\] \end{theorem} The forward implication follows by induction on the given derivation. The reverse implication is immediate from the rule which allows to fall back on Guarded HOL in relational proofs. (Rule \rname{SUB} in the appendix). The full proof is in the appendix. The consequence of this theorem is that the syntax-directed, relational proof system we have built on top of Guarded HOL does not lose expressiveness. The intended semantics of a judgement $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi}$ is that, for every valuation $\delta \models \Delta$, $\gamma\models \Gamma$, if $\sem{\Sigma}(\delta)$ and $\sem{\Psi}(\delta,\gamma)$, then $$\sem{\phi}(\delta,\gamma[\res\ltag \ot \sem{t_1}(\delta,\gamma), \res\rtag\ot\sem{t_2}(\delta,\gamma)])$$ Since Guarded HOL is sound with respect to its semantics in the topos of trees, and our relational proof system is equivalent to Guarded HOL, we obtain that our relational proof system is also sound in the topos of trees. \begin{corollary}[Soundness and consistency]\label{cor:rhol:sound} If $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_2}{t_2}{\sigma_2}{\phi}$, then for every valuation $\delta \models \Delta$, $\gamma\models\Gamma$: \[\begin{array}{c} \sem{\Delta \vdash \square \Sigma}(\delta) \wedge \sem{\Delta \mid \Gamma \vdash \Psi}(\delta,\gamma) \Rightarrow \\ \sem{\Delta \mid \Gamma,\res\ltag:\sigma_1, \res\ltag:\sigma_2 \vdash \phi} (\delta, \gamma[\res\ltag \ot \sem{\Delta \mid \Gamma \vdash t_1}(\delta,\gamma)][\res\rtag \ot \sem{\Delta \mid \Gamma \vdash t_2}(\delta,\gamma)]) \end{array}\] In particular, there is no proof of $\jgrhol{\Delta}{\emptyset}{\Gamma}{\emptyset}{t_1}{\sigma_1}{t_2}{\sigma_2}{\bot}$. \end{corollary} \subsection{Shift couplings revisited} \label{sec:examples} We give further details on how to prove the example with shift couplings from Section~\ref{sec:proba-ex}. (Additional examples of relational reasoning on non-probabilistic streams can be found in the appendix.) Recall the step functions:$$\begin{array}{rcl} \operatorname{step} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\munit{z+x}} \\ \operatorname{lstep2} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\mlet{b}{\mathcal{U}_{\{0,1\}}}{\munit{x+2*z*b}}} \end{array}$$ We axiomatize the predicate $\All_{2,1}$, which relates the element at position $2i$ in one stream to the element at position $i$ in another stream, as follows. \[\begin{array}{l} \forall x_1 x_2 xs_1 xs_2 y_1. \phi\subst{x_1}{z_1}\subst{x_2}{z_2} \Rightarrow \\ \quad \later[ys_1\ot xs_1]{\later[zs_1 \ot ys_1, ys_2 \ot xs_2]{\All_{2,1}(zs_1, ys_2, z_1.z_2.\phi)}} \Rightarrow \\ \quad\quad \All_{2,1}(\cons{x_1}{\cons{y_1}{xs_1}},\cons{x_2}{xs_2}, z_1.z_2. \phi) \end{array} \] In fact, we can assume that, in general, we have a family of $\All_{m_1, m_2}$ predicates relating two streams at positions $m_1\cdot i$ and $m_2\cdot i$ for every $i$. We can now express the existence of a shift coupling by the statement: {\small \[p_1 = p_2 \vdash \markov(p_1, \operatorname{step}) \sim \markov(p_2, \operatorname{lstep2}) \mid \diamond_{\left[\substack{ y_1 \ot \res\ltag\\ y_2 \ot \res\rtag}\right]}\All_{2,1}(y_1,y_2,z_1.z_2.z_1=z_2) \]} For the proof, we need to introduce an asynchronous rule for Markov chains: {\small\[ \inferrule*[right=\sf Markov-2-1] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi} \\\\ \begin{array}{c} \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{(\lambda x_1. \mlet{x'_1}{h_1\ x_1}{h_1\ x'_1})}{C_1\to\Distr(C_1)}{h_2}{C_2\to \Distr(C_2)} {\\ \forall x_1 x_2. \phi\subst{z_1}{x_1}\subst{z_2}{x_2} \Rightarrow \diamond_{[ z_1 \ot \res\ltag\ x_1, z_2 \ot \res\rtag\ x_2]}\phi} \end{array}} {\begin{array}{c}\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\markov(t_1,h_1)}{\Distr(\Str{C_1})}{\markov(t_2,h_2)}{\Distr(\Str{C_2})} {\\ \diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]}\All_{2,1}(y_1,y_2,z_1.z_2.\phi)} \end{array}} \]} This asynchronous rule for Markov chains shares the motivations of the rule for loops proposed in~\cite{BartheGHS17}. Note that one can define a rule \rname{Markov-m-n} for arbitrary $m$ and $n$ to prove a judgement of the form $\All_{m,n}$ on two Markov chains. We show the proof of the shift coupling. By equational reasoning, we get: \[ \begin{array}{rl} & \lambda x_1. \mlet{x'_1}{h_1\ x_1}{h_1\ x'_1}\; \equiv \; \lambda x_1. \mlet{z_1}{\unif{\{-1,1\}}} {h_1\ (z_1 + x_1)} \; \equiv \\ &\equiv \; \lambda x_1. \mlet{z_1}{\unif{\{-1,1\}}} {\mlet{z'_1}{\unif{\{-1,1\}}} {\munit{z'_1 + z_1 + x'_1}}} \end{array} \] and the only interesting premise of \rname{Markov-2-1} is: \[\left.\begin{array}{rl} \lambda x_1. &\mlet{z_1}{\unif{\{-1,1\}}} {\\ &\mlet{z'_1}{\unif{\{-1,1\}}} {\\ &\munit{z'_1 + z_1 + x'_1}}} \end{array} \sim \begin{array}{rl} \lambda x_2. &\mlet{z_2}{\unif{\{-1,1\}}} {\\ &\mlet{b_2}{\unif{\{1,0\}}} {\\ &\munit{x_2+2*b_2*z_2}}} \end{array} \;\right|\; \begin{array}{c} \forall x_1 x_2. x_1 = x_2 \Rightarrow \\ \res\ltag\ x_1 \stackrel{\diamond}{=} \res\rtag\ x_2 \end{array}\] Couplings between $z_1$ and $z_2$ and between $z_1'$ and $b_2$ can be found by simple computations. This completes the proof. \section{Related work} \label{sec:rw} Our probabilistic guarded $\lambda$-calculus and the associated logic Guarded HOL build on top of the guarded $\lambda$-calculus and its internal logic~\cite{CBGB16}. The guarded $\lambda$-calculus has been extended to guarded dependent type theory~\cite{BGCMB16}, which can be understood as a theory of guarded refinement types and as a foundation for proof assistants based on guarded type theory. These systems do not reason about probabilities, and do not support syntax-directed (relational) reasoning, both of which we support. Relational models for higher-order programming languages are often defined using logical relations. \cite{PlotkinA93} showed how to use second-order logic to define and reason about logical relations for the second-order lambda calculus. Recent work has extended this approach to logical relations for higher-order programming languages with computational effects such as nontermination, general references, and concurrency~\cite{DreyerAB11,caresl,ipm,Krogh-Jespersen17}. The logics used in \emph{loc. cit.} are related to our work in two ways: (1) the logics in \emph{loc. cit.} make use of the later modality for reasoning about recursion, and (2) the models of the logics in \emph{loc. cit.} can in fact be defined using guarded type theory. Our work is more closely related to Relational Higher Order Logic~\cite{ABGGS17}, which applies the idea of logic-enriched type theories~\cite{AczelG00,AczelG06} to a relational setting. There exist alternative approaches for reasoning about relational properties of higher-order programs; for instance,~\cite{GrimmMFHMPRSB17} have recently proposed to use monadic reification for reducing relational verification of $F^*$ to proof obligations in higher-order logic. A series of work develops reasoning methods for probabilistic higher-order programs for different variations of the lambda calculus. One line of work has focused on operationally-based techniques for reasoning about contextual equivalence of programs. The methods are based on probabilistic bisimulations~\cite{DBLP:conf/esop/CrubilleL14,DBLP:conf/popl/SangiorgiV16} or on logical relations~\cite{DBLP:conf/fossacs/BizjakB15}. Most of these approaches have been developed for languages with discrete distributions, but recently there has also been work on languages with continuous distributions~\cite{DBLP:conf/icfp/BorgstromLGS16,DBLP:conf/esop/CulpepperC17}. Another line of work has focused on denotational models, starting with the seminal work in~\cite{Jones:powerdomain-evaluations}. Recent work includes support for relational reasoning about equivalence of programs with continuous distributions for a total programming language~\cite{DBLP:conf/lics/StatonYWHK16}. Our approach is most closely related to prior work based on relational refinement types for higher-order probabilistic programs. These were initially considered by~\cite{BFGSSZ14} for a stateful fragment of $F^*$, and later by~\cite{BGGHRS15,BFGGGHS16} for a pure language. Both systems are specialized to building probabilistic couplings; however, the latter support approximate probabilistic couplings, which yield a natural interpretation of differential privacy~\cite{DR14}, both in its vanilla and approximate forms (i.e.\, $\epsilon$- and $(\epsilon,\delta)$-privacy). Technically, approximate couplings are modelled as a graded monad, where the index of the monad tracks the privacy budget ($\epsilon$ or $(\epsilon,\delta)$). Both systems are strictly syntax-directed, and cannot reason about computations that have different types or syntactic structures, while our system can. \section{Conclusion} \label{sec:conclusion} We have developed a probabilistic extension of the (simply typed) guarded $\lambda$-calculus, and proposed a syntax-directed proof system for relational verification. Moreover, we have verified a series of examples that are beyond the reach of prior work. Finally, we have proved the soundness of the proof system with respect to the topos of trees. There are several natural directions for future work. One first direction is to enhance the expressiveness of the underlying simply typed language. For instance, it would be interesting to introduce clock variables and some type dependency as in~\cite{BGCMB16}, and extend the proof system accordingly. This would allow us, for example, to type the function taking the $n$-th element of a \emph{guarded} stream, which cannot be done in the current system. Another exciting direction is to consider approximate couplings, as in~\cite{BGGHRS15,BFGGGHS16}, and to develop differential privacy for infinite streams---preliminary work in this direction, such as~\cite{KellarisPXP14}, considers very large lists, but not arbitrary streams. A final direction would be to extend our approach to continuous distributions to support other application domains. \subsubsection*{Acknowledgments.} We would like to thank the anonymous reviewers for their time and their helpful input. This research was supported in part by the ModuRes Sapere Aude Advanced Grant from The Danish Council for Independent Research for the Natural Sciences (FNU), by a research grant (12386, Guarded Homotopy Type Theory) from the VILLUM foundation, and by NSF under grant 1718220. \bibliography{refs} \appendix \newpage \section{Additional proof rules} \begin{figure*}[!htb] \small \begin{mathpar} \infer[\sf ABS] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\lambda x_1:\tau_1. t_1}{\tau_1 \to \sigma_1}{\lambda x_2:\tau_2. t_2}{\tau_2\to \sigma_2}{\forall x_1,x_2. \phi' \Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}\subst{\res\rtag}{\res\rtag\ x_2}}} {\jgrhol{\Delta}{\Sigma}{\Gamma,x_1:\tau_1,x_2:\tau_2}{\Psi,\phi'}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}} \and \infer[\sf APP]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1 u_1}{\sigma_1}{t_2 u_2}{\sigma_2}{\phi\subst{x_1}{u_1}\subst{x_2}{u_2}}} {\begin{array}{c} \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\tau_1\to \sigma_1}{t_2}{\tau_2\to \sigma_2}{ \forall x_1,x_2. \phi'\subst{\res\ltag}{x_1}\subst{\res\rtag}{x_2}\Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}\subst{\res\rtag}{\res\rtag\ x_2}}\\ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\tau_1}{u_2}{\tau_2}{ \phi'} \end{array}} \and \infer[\sf VAR]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{\sigma_1}{x_2}{\sigma_2}{\phi}} {\jlc{\Delta\mid\Gamma}{x_1}{\sigma_1} & \jlc{\Delta\mid\Gamma}{x_2}{\sigma_2} & \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res\ltag}{x_1}\subst{\res\rtag}{x_2}}} \and \infer[\sf SUB]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi'} & \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash_{\sf GHOL} \phi'\defsubst{t_1}{t_2} \Rightarrow \phi\defsubst{t_1}{t_2}} \and \infer[\sf UHOL-L] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{\phi\defsubst{\res}{t_2}} \\ \jlc{\Delta\mid\Gamma}{t_2}{\sigma_2}} \and \infer[\sf ABS{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\lambda x_1:\tau_1. t_1}{\tau_1 \to \sigma_1}{t_2}{\sigma_2}{\forall x_1. \phi' \Rightarrow \phi \subst{\res\ltag}{\res\ltag\ x_1}}} {\jgrhol{\Delta}{\Sigma}{\Gamma,x_1:\tau_1}{\Psi, \phi'}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}} \and \infer[\sf APP{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1 u_1}{\sigma_1}{u_2}{\sigma_2}{\phi\subst{x_1}{u_1}}} {\begin{array}{c} \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\tau_1\to \sigma_1}{u_2}{\sigma_2}{\forall x_1. \phi'\subst{\res\ltag}{x_1} \Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}}\\ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\sigma_1}{\phi'} \end{array}} \and \infer[\sf VAR{-}L] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{\sigma_1}{t_2}{\sigma_2}{\phi}} {\phi\subst{\res\ltag}{x_1} \in \Psi & \res\rtag\not\in\ FV(\phi) & \jlc{\Delta\mid\Gamma}{t_2}{\sigma_2}} \and \inferrule*[Right=\sf Equiv] {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t'_1}{A_1}{t'_2}{A_2}{\Phi} \\ t_1 \equiv t_1' \\ t_2 \equiv t_2' \\ \Delta \mid \Gamma \vdash t_1 : A_1 \\ \Delta \mid \Gamma \vdash t_2 : A_2} {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi}} \end{mathpar} \caption{Selected RHOL rules}\label{fig:sel-rhol} \end{figure*} \section{Denotational semantics} \subsection{Types and terms in context} The meaning of terms is given by the denotational model in the category $\trees$ of presheaves over $\omega$, the first infinite ordinal. This category $\trees$ is also known as the \emph{topos of trees}~\cite{Birkedal-et-al:topos-of-trees}. In previous work~\cite{CBGB16} it was shown how to model most of the constructions of the guarded lambda calculus and the associated logic, with the notable exception of the probabilistic features. Below we give an elementary and self-contained presentation of the semantics. Concretely, objects $X$ of $\trees$ are families of sets $X_i$ indexed over $\nat$ together with functions $r_n^X : X_{n+1} \to X_n$. These are called \emph{restriction functions}. We will write simply $r_n$ if $X$ is clear from the context. Moreover if $x \in X_i$ and $j \leq i$ we will write $x\restriction_j$ for the element $r_j(\cdots (r_{i-1}(x))\cdots) \in X_j$. Morphisms $X \to Y$ are families of functions $\alpha_n : X_n \to Y_n$ commuting with restriction functions in the sense of $r_n^Y \circ \alpha_{n+1} = \alpha_n \circ r_n^X$. One can see the restriction function $r_n : X_{n+1} \to X_n$ as mapping elements of $X_{n+1}$ to their approximations at time $n$. Semantics of types can be found on \autoref{fig:sem-types}, where $G\left(\sem{A}\right)$ consists of sequences $\{x_n\}_{n \in \nat}$ such that $x_i \in \sem{A}_i$ and $r_i(x_{i+1}) = x_i$ for all $i$, i.e., $\square\sem{A}$ is the set of so-called global sections of $\sem{A}$. The semantics of a dual context $\Delta \mid \Gamma$ is given as the product of types in $\Delta$ and $\Gamma$, except that we implicitly add $\square$ in front of every type in $\Delta$. In the particular case when both contexts are empty, the semantics of the dual context correspond to the terminal object $1$, which is the singleton set $\{\ast\}$ at each stage. A term in context $\Delta \mid \Gamma \vdash t : \tau$ is interpreted as a family of functions $\sem{t}_n : \sem{\Delta \mid \Gamma}_n \to \sem{\tau}_n$ commuting with restriction functions of $\sem{\Delta \mid \Gamma}$ and $\sem{\tau}$. Semantics of products, coproducts, and natural numbers is pointwise as in sets, so we omit writing it. The cases for the other constructs are in \autoref{fig:sem-glc} where $\mathsf{munit}$ and $\mathsf{mlet}$ are the standard unit and bind operations on discrete probabilities, i.e.\, \begin{align*} \begin{array}{rcl} \munit{c} & = & \lambda y.~\carac{c=y} \\ \mathsf{mlet}~x = \mu~\mathsf{in}~M & = & \lambda y.~\sum_{c\in C} \mu(c) \cdot M(c)(y) \end{array} \end{align*} The functions $\pi_0$ and $\pi_1$ are the first and second projections, respectively. \begin{figure*}[!htb] \begin{align*} \sem{b} &\defeq \text{ chosen object of } \trees\\ \sem{\nat} &\defeq \nat \xleftarrow{id} \nat \xleftarrow{id} \nat \xleftarrow{id} \cdots\\ \sem{A\times B} &\defeq \sem{A}_0\times\sem{B}_0 \xleftarrow{r_0 \times r_0} \sem{A}_1\times\sem{B}_1 \xleftarrow{r_1 \times r_1} \sem{A}_2\times\sem{B}_2 \xleftarrow{r_2 \times r_2} \cdots \\ \sem{A\to B} &\defeq \left(\sem{B}^{\sem{A}}\right)_0 \xleftarrow{\pi} \left(\sem{B}^{\sem{A}}\right)_1 \xleftarrow{\pi} \left(\sem{B}^{\sem{A}}\right)_2 \xleftarrow{\pi} \cdots \\ \sem{\Str{A}} &\defeq \sem{A}_0 \times \{*\} \xleftarrow{r_0 \times !} \sem{A}_1 \times \left(\sem{A}_0 \times \{*\}\right) \xleftarrow{r_1 \times r_0 \times !} \sem{A}_2\times\left(\sem{A}_1\times\left(\sem{A}_0 \times \{*\}\right)\right) \leftarrow \cdots \\ \sem{\later A} &\defeq \{*\} \xleftarrow{!} \sem{A}_0 \xleftarrow{r_0} \sem{A}_1 \xleftarrow{r_1} \cdots \\ \sem{\square A} &\defeq G(\sem{A}) \xleftarrow{id} G(\sem{A}) \xleftarrow{id} \cdots\\ \sem{\Distr(C)} &\defeq \Distr(\sem{C}_0) \stackrel{\Distr(r_0)}{\longleftarrow} \Distr(\sem{C}_1) \stackrel{\Distr(r_1)}{\longleftarrow} \Distr(\sem{C}_2) \stackrel{\Distr(r_2)}{\longleftarrow} \ldots \end{align*} \caption{Semantics of types in the topos of trees} \label{fig:sem-types} \end{figure*} \begin{figure*}[!htb] \begin{align*} &\sem{\Delta \mid \Gamma \vdash \lambda x:A. t : A \to B}_i(\delta,\gamma) \defeq (f_0,\ldots,f_i)\\ &\quad\quad\text{ where } f_i(x) = \sem{\Delta\mid \Gamma, x : A \vdash t : B}\left(\delta,(\gamma\restriction_i, x)\right)\\ &\sem{\Delta \mid \Gamma \vdash t_1\ t_2 : B}_i(\delta,\gamma) \defeq f_i\left(\sem{\Delta \mid\Gamma \vdash t_2 : A}(\delta,\gamma)\right)\\ &\quad\quad\text{ where } \sem{\Delta \mid \Gamma \vdash t_1 : A \to B}_i(\delta,\gamma) = (f_0,\ldots,f_i)\\ &\sem{\Delta \mid \Gamma \vdash \later[x_1\ot t_1, \dots, x_n\ot t_n]t : \later A}_0(\delta,\gamma) \defeq *\\ &\sem{\Delta \mid \Gamma \vdash \later[x_1\ot t_1, \dots, x_n\ot t_n]t : \later A}_{i+1}(\delta,\gamma) \defeq \\ &\quad\quad\sem{\Delta \mid \Gamma, \{x_k : A_k\}_{k=1}^n \vdash t : A}_i\left(\delta, (\gamma\restriction_i, \left\{\sem{\Delta \mid \Gamma \vdash t_k : \later A_k}_{i+1}\right\}_{k=1}^n(\delta,\gamma))\right)\\ &\sem{\Delta \mid \Gamma \vdash \prev t : A}_i(\delta,\gamma) \defeq \sem{\Delta \mid \cdot \vdash t : \later A}_{i+1}(\delta)\\ &\sem{\Delta \mid \Gamma \vdash \boxx t : \square A}_i(\delta,\gamma) \defeq \{\sem{\Delta \mid \cdot \vdash t: A}_j(\delta)\}_{j=0}^{\infty}\\ &\sem{\Delta \mid \Gamma \vdash \letbox{x}{u}{t} : A}_i(\delta,\gamma) \defeq \\ &\quad\quad\sem{\Delta, x : B \mid \Gamma \vdash t : A}_i\left(\left(\delta,\sem{\Delta \mid \Gamma \vdash u : \square B}_i(\delta,\gamma)\right),\gamma\right)\\ &\sem{\Delta \mid \Gamma \vdash \letconst{x}{u}{t} : A}_i(\delta,\gamma)\defeq\\ &\quad\quad\sem{\Delta, x : B \mid \Gamma \vdash t : A}_i\left(\left(\delta,\varepsilon^{-1}_i\left(\sem{\Delta \mid \Gamma \vdash u : \square B}_i(\delta,\gamma)\right)\right),\gamma\right)\\ &\sem{\Delta \mid \Gamma \vdash \hd t : A}_i(\delta,\gamma) \defeq \pi_0\left(\sem{\Delta \mid \Gamma \vdash t : \Str{A}}_i(\delta,\gamma)\right)\\ &\sem{\Delta \mid \Gamma \vdash \tl t : \later{\Str{A}}}_i(\delta,\gamma) \defeq \pi_1\left(\sem{\Delta \mid \Gamma \vdash t : \Str{A}}_i(\delta,\gamma)\right)\\ &\sem{\Delta \mid \Gamma \vdash \cons{t}{u} : \Str{A}}_i(\delta,\gamma) \defeq \left(\sem{\Delta \mid \Gamma \vdash t : A}_i(\delta,\gamma), \sem{\Delta \mid \Gamma \vdash u : \Str{A}}_i(\delta,\gamma)\right)\\ &\sem{\Delta \mid \Gamma \vdash \munit{t} : \Distr(C)}_i (\delta,\gamma) \defeq \munit{\sem{\Delta \mid \Gamma \vdash t:C}_i (\delta,\gamma)} \\ &\sem{\Delta \mid \Gamma \vdash \mlet{x}{t}{u} : \Distr(C)}_i (\delta,\gamma) \defeq \mathsf{mlet}~v= \sem{\Delta \mid \Gamma \vdash t : \Distr(D)}_i (\delta,\gamma)~\mathsf{in} \\ \mbox{}\hspace{0.5cm} &\quad\quad\sem{\Delta \mid \Gamma,x:D \vdash u : \Distr(C)}_i (\delta,\gamma[x:=v]) \end{align*} \caption{Semantics for the Guarded $\lambda$-calculus} \label{fig:sem-glc} \end{figure*} \subsection{Equational theory of the calculus} The denotational semantics validates the following equational theory in addition to the standard equational theory of the simply typed lambda calculus with sums and natural numbers.\\ \textbf{Rules for fixed points, always modality and streams} \begin{align*} \begin{split} \fix{f}{t} \quad &\equiv \quad t\subst{f}{\later(\fix{f}{t})} \\ \prev{(\later t)} \quad &\equiv \quad t \\ \letbox{x}{(\boxx{u})}{t} \quad &\equiv \quad t\subst{x}{u} \\ \letconst{x}{u}{t} \quad&\equiv\quad t\subst{x}{u} \\ \end{split} \begin{split} \hd\ (\cons{x}{xs}) \quad&\equiv\quad x \\ \tl\ (\cons{x}{xs}) \quad&\equiv\quad xs \\ \cons{\hd t}{\tl t} \quad&\equiv\quad t \end{split} \end{align*} \textbf{Rules for delayed substitutions} \begin{align*} \latern[\xi\hrt{x\gets t}]{u} &\equiv \latern[\xi]{u} & & \text{ if } x \text{ not in } u\\ \latern[\xi\hrt{x\gets t, y \gets s}\xi']{u} &\equiv\latern[\xi\hrt{y\gets s, x \gets t}\xi']{u}\\ \latern[\xi\hrt{x\gets \latern[\xi]{t}}]{u} &\equiv \latern[\xi]{\left(u\subst{x}{t}\right)}\\ \latern[\hrt{x\gets t}]x &\equiv t \end{align*} \textbf{Monad laws for distributions} \begin{align*} \mlet{x}{\munit{t}}{u} &\equiv u[t/x]\\ \mlet{x}{t}{\munit{x}} &\equiv t\\ \mlet{x_2}{(\mlet{x_1}{t_1}{t_2})}{u} &\equiv \mlet{x_1}{t_1}{(\mlet{x_2}{t_2}{u})} \end{align*} In particular, notice that fix does not reduce as usual, but instead the whole term is delayed before the substitution is performed. \subsection{Logical judgements} The cases for the semantics of the judgement $\Delta \mid \Gamma \vdash \phi$ of the non-probabilistic fragment are as follows (we omit writing the contexts if they are clear): \begin{align*} \sem{\top}_i &\defeq \sem{\Delta \mid \Gamma}_i\\ \sem{\phi \wedge \psi}_i &\defeq \sem{\phi}_i \cap \sem{\psi}_i\\ \sem{\phi \vee \psi}_i &\defeq \sem{\phi}_i \cup \sem{\psi}_i\\ \sem{\phi \To \psi}_i &\defeq \left\{ x \isetsep \forall j \leq i, x\restriction_j \in \sem{\phi}_j \To x\restriction_j \in \sem{\psi}_j\right\}\\ \sem{\forall x:A. \phi}_i &\defeq \left\{(\delta,\gamma) \isetsep \forall j \leq i, \forall x \in \sem{A}_j, \left(\delta, \left(\gamma\restriction_j\right), x\right) \in \sem{\phi}\right\}\\ \sem{\later[x_1 \ot t_1, \dots, x_n \ot t_n]\phi}_i &\defeq \left\{ (\delta, \gamma) \isetsep i > 0 \To \left(\delta,\gamma\restriction_{i-1}, \left\{\sem{t_k}_{i}(\delta,\gamma)\right\}_{k=1}^n\right) \in \sem{\phi}_{i-1}\right\}\\ \sem{\square \phi}_i &\defeq \left\{ x \isetsep \forall j, x \in \sem{\phi}_j\right\} \end{align*} \section{Additional background} One consequence of Strassen's theorem is that couplings are closed under convex combinations. \begin{lemma}[Convex combinations of couplings]\label{lem:convex-combination-of-couplings} Let $(\mu_i)_{i\in I}$ and $(\nu_i)_{i\in I}$ bet two families of distributions on $C_1$ and $C_2$ respectively, and let $(p_i)_{i\in I} \in [0,1]$ such that $\sum_{i\in I} p_i=1$. If $\coupl{\mu_i}{\nu_i}{R}$ for all $i\in I$ then $\coupl{(\sum_{i\in I} p_i \mu_i)}{(\sum_{i\in I} p_i \nu_i)}{R}$, where the convex combination $\sum_{i\in I} p_i \mu_i$ is defined by the clause $(\sum_{i\in I} p_i \mu_i)(x)=\sum_{i\in I} p_i \mu_i(x)$. \end{lemma} One obtains an asymmetric version of the lemma by observing that if $\mu_i=\mu$ for every $i\in I$, then $(\sum_{i\in I} p_i \mu_i)=\mu$. One can also show that couplings are closed under relation composition. \begin{lemma}[Couplings for relation composition] Let $\mu_1\in\Distr(C_1)$, $\mu_2\in\Distr(C_2)$, $\mu_3 \in\Distr(C_3)$. Moreover, let $R\subseteq C_1\times C_2$ and $S\subseteq C_2\times C_3$. If $\coupl{\mu_1}{\mu_2}{R}$ and $\coupl{\mu_2}{\mu_3}{R}$ then $\coupl{\mu_1}{\mu_3}{R}$. \end{lemma} \section{Proofs of the theorems} \subsection{Proof of Theorem~\ref{thm:sound-ghol}} \label{sec:soundness-of-ghol} The semantics of the guarded higher-order logic without the probabilistic fragment has been explained in previous work~\cite{Birkedal-et-al:topos-of-trees,CBGB16}. Thus we focus on showing soundness of the additional rules for the diamond modality, which will be useful for proving soundness of the relational proof system. Moreover we only describe soundness for the binary diamond modality, the soundness of the rules for the unary modality being entirely analogous. \noindent\textbf{Soundness of the rule \textsf{MONO2}} \begin{mathpar} \infer[\sf MONO2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\gets t_1, x_2\gets t_2]}\psi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\gets t_1, x_2 \gets t_2]}\phi} \and \jghol{\Delta}{\Sigma}{\Gamma, x_1 : C_1, x_2 : C_2}{\Psi, \phi}{\psi} } \end{mathpar} Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$. Then from the first premise we have $(\delta, \gamma) \in \sem{\diamond_{[x_1\gets t_1, x_2 \gets t_2]}\phi}_n$ and thus there exists an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}$ coupling for the distributions $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$. But since $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$ we have from the second premise of the rule that \begin{align*} \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\} \subseteq \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\} \end{align*} and thus any $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}$ coupling is also an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\}$ coupling, which means there exists an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\}$ coupling for $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$ as required. \noindent\textbf{Soundness of the rule \textsf{UNIT2}} \begin{mathpar} \infer[\sf UNIT2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[x_1\leftarrow \munit{t_1}, x_2\leftarrow \munit{t_2}]} \phi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \phi[t_1/x_1][t_2/x_2]}} \end{mathpar} Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$. We need to show the existence of a \begin{align*} \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\} \end{align*} coupling for the point-mass distributions concentrated at $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$. The premise of the rule establishes the membership \begin{align*} \left(\sem{t_1}_n(\delta,\gamma), \sem{t_2}_n(\delta,\gamma)\right) \in \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\} \end{align*} and thus the point-mass distribution concentrated at $\left(\sem{t_1}_n(\delta,\gamma), \sem{t_2}_n(\delta,\gamma)\right)$ is a required coupling. \noindent\textbf{Soundess of the rule \textsf{MLET2}} \begin{mathpar} \infer[\sf MLET2] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow \mlet{x_2}{t_2}{t'_2}]} \psi}}{ \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]} \phi} \and \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi}{ \diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}} \end{mathpar} Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$. Then from the first premise we have that there exists an \begin{align*} \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\} \end{align*} coupling for the distributions $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$. From the second premise we get that for every $v, u$ such that $(\delta, \gamma, v, u) \in \sem{\phi}_n$ there exists an \begin{align*} \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\} \end{align*} coupling for $\sem{t_1'}_n(\delta,\gamma, v)$ and $\sem{t_2'}_n(\delta,\gamma, u)$. Since $x_1$ and $x_2$ are fresh for $\psi$ the relation \begin{align*} \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\} \end{align*} is independent of $v, u$. Thus Lemma~\ref{lem:sequential-composition-of-couplings} instantiated with \begin{align*} \mu_1 &= \sem{t_1}_n(\delta,\gamma)\\ \mu_2 &= \sem{t_2}_n(\delta,\gamma)\\ M_1 &= \sem{t_1'}_n(\delta,\gamma, -)\\ M_2 &= \sem{t_2'}_n(\delta,\gamma, -)\\ R &= \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}\\ S &= \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\} \end{align*} concludes the proof. \noindent\textbf{Soundness of the rule \textsf{MLET-L}} \begin{mathpar} \infer[\sf MLET-L] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow t'_2]} \psi}}{ \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1]}\phi} \and \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi}{ \diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}} \end{mathpar} Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$. Then from the first premise we have that the support of the distribution $\sem{t_1}_n(\delta,\gamma)$ is included in \begin{align*} \left\{v \isetsep (\delta, \gamma, v) \in \sem{\phi}_n\right\}. \end{align*} From the second premise we get that for every $v$ such that $(\delta, \gamma, v) \in \sem{\phi}_n$ there exists an \begin{align*} \left\{(v', u') \isetsep (\delta, \gamma, v, v', u') \in \sem{\psi}_n\right\} \end{align*} coupling for $\sem{t_1'}_n(\delta,\gamma, v)$ and $\sem{t_2'}_n(\delta,\gamma)$. Since $x_1$ is fresh for $\psi$ the relation \begin{align*} R \defeq \left\{(v', u') \isetsep (\delta, \gamma, v, v', u') \in \sem{\psi}_n\right\} \end{align*} is independent of $v$. Let $\mathcal{I} = \left\{v \isetsep (\delta, \gamma, v) \in \sem{\phi}_n\right\}$ and for any $v \in \mathcal{I}$ let $p_v = \sem{t_1}_n(\delta,\gamma)(v)$, $\mu_v = \sem{t_1'}_n(\delta,\gamma,v)$, and $\nu_v = \sem{t_2'}_n(\delta,\gamma)$. Then we have $\sum_{v \in \mathcal{I}} p_v = 1$ from the first premise of the rule and $\coupling{R}{\mu_v}{\nu_v}$ for all $v \in \mathcal{I}$ from the second premise. Lemma~\ref{lem:convex-combination-of-couplings} concludes the proof. \subsection{Proof of Theorem \ref{thm:equiv-rhol-hol}} The inverse implication follows immediately from the \rname{SUB} rule and the fact that we can always prove a judgement of the shape \[ \jgrhol{\Gamma}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\top} \] for well-typed $t_1$ and $t_2$. We will prove the direct implication by induction on the derivation. We will just prove the two-sided rules. The proofs for the one sided rule are similar. \\ \\ \noindent {\bf Case.} $\small\inferrule*[Right=\sf Next] {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_2}{\Psi, \Phi'\defsubst{x_1}{x_2}}{t_1}{A_1}{t_2}{A_2}{\Phi} \\ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{A_1}}{u_2}{\later{A_2}}{\dsubst{\res\ltag, \res\rtag}{\res\ltag, \res\rtag}{\Phi'}}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi} {\nextt{x_1 \leftarrow u_1}{t_1}}{\later{A_1}} {\nextt{x_2 \leftarrow u_2}{t_2}}{\later{A_2}} {\triangleright[x_1\leftarrow u_1,x_2\leftarrow x_2,\res\ltag\leftarrow\res\ltag,\res\rtag\leftarrow\res\rtag]\Phi}}$ \\ \\ \noindent By I.H. $\jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_n}{\Psi, \Phi'\defsubst{x_1}{x_2}}{\Phi\defsubst{t_1}{t_2}}$, \hfill{}(H1) \\ and $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[\res\ltag \ot u_1, \res\rtag \ot u_2]{\Phi'}}$ \hfill{}(H2) \\ To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2, \res\ltag \ot \nextt{x_1 \leftarrow u_1}{t_1} , \res\rtag\ot \nextt{x_2 \leftarrow u_2}{t_2}]{\Phi}}$. \hfill{}(G) \\ By \rname{CONV} we can change the goal (G) into \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2]{\Phi\defsubst{t_1}{t_2}}}$ \hfill{}(G') \\ and (H2) into: \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2]{\Phi'\defsubst{x_1}{x_2}}}$ \hfill{}(H2') \\ Finally, by applying \rname{$\later_{App}$} to (H1) and (H2) we get (G') \\ \noindent {\bf Case.} $\small \inferrule*[Right=\sf Prev] {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{\later A_2}{\dsubst{\res\ltag,\res\rtag}{\res\ltag,\res\rtag}{\Phi}}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t_1}}{A_1}{\prev{t_2}}{A_2}{\Phi}}$ \\ \\ \noindent We just apply \rname{$\later_E$}. \\ \noindent {\bf Case.} $\small\inferrule*[Right=\sf Box] {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{A_1}{t_2}{A_2}{\Phi}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t_1}}{\square A_1}{\boxx{t_2}}{\square A_2} {\square \Phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}}}$ \\ \\ \noindent By I.H. $\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\Phi\defsubst{t_1}{t_2}}$ \\ To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{\letbox{x_1}{\boxx{t_1}}{x_1}}{\letbox{x_2}{\boxx{t_2}}{x_2}}}$ \\ By \rname{CONV} we can change the goal into: \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{t_1}{t_2}}$ \\ And then we can prove it by \rname{$\square_I$}. \\ \noindent {\bf Case.}$\small\inferrule*[Right=\sf LetBox] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{u_2}{\square B_2}{\square \Phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}} \\ \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi'}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{\letbox{x_2}{u_2}{t_2}}{A_2}{\Phi'}}$ \\ \\ \noindent By I.H. $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}$ \hfill{}(H1) \\ and $\jghol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{\Phi'\defsubst{t_1}{t_2}}$. \hfill{}(H2) \\ To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi'\defsubst{\letbox{x_1}{u_1}{t_1}}{\letbox{x_2}{u_2}{t_2}}}$ \hfill{}(G) \\ We instantiate (H2) into: \[\begin{array}{c} \jghol{\Delta}{\Sigma, \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}{\Gamma}{\Psi} {\\ \Phi'\defsubst{t_1\subst{x_1}{\letbox{x_1}{u_1}{x_1}}}{t_2\subst{x_2}{\letbox{x_2}{u_2}{x_2}}}} \end{array}\] And by the equality $t\subst{x}{\letbox{x}{u}{x}} \equiv \letbox{x}{u}{t}$, we get: \[\begin{array}{c} \jghol{\Delta}{\Sigma, \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}{\Gamma}{\Psi} {\\ \Phi'\defsubst{\letbox{x_1}{u_1}{t_1}}{\letbox{x_2}{u_2}{t_2}}} \end{array} \] and then, by applying \rname{$\square_E$} to (H1) and the previous judgement we get (G). \\ \noindent {\bf Case.} $\small \inferrule*[Right=\sf LetConst] {B_1,B_2,\Phi\ \text{constant} \\ FV(\Phi)\cap FV(\Gamma) = \emptyset \\ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{u_2}{B_2}{\Phi} \\ \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi'}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{\letconst{x_2}{u_2}{t_2}}{A_2}{\Phi'}}$ \\ \\ \noindent By I.H. $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi\subst{u_1}{u_2}}$, \hfill{}(H1) \\ and $\jghol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{\Phi'\defsubst{t_1}{t_2}}$. \hfill{}(H2) \\ To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi'\defsubst{\letconst{x_1}{u_1}{t_1}}{\letconst{x_2}{u_2}{t_2}}}$ \hfill{}(G) \\ From (H1) and the fact that $\Phi, B_1$ and $B_2$ are constant, we get: \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\Phi\defsubst{u_1}{u_2}}$ \\ The rest of the prove is analogous to the previous case. \\ \noindent {\bf Case.} $\small \inferrule*[Right=\sf Fix] {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi, \dsubst{\res\ltag,\res\rtag}{f_1,f_2}{\Phi}} {t_1}{A_1}{t_2}{A_2}{\Phi}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{fix f_1. t_1}{A_1}{fix f_2. t_2}{A_2}{\Phi}}$ \\ \\ \noindent By I.H. $\jghol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi,\later[\res\ltag\ot f_1, \res\rtag\ot f_2]{\Phi}}{\Phi\defsubst{t_1}{t_2}}$. \\ To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi\subst{\fix{f_1}{t_1}}{fix{f_2}{t_2}}}$ \\ Instantiating the I.H. with $f_1 = \later \fix{f_1}{t_1}$ and $f_2 = \later \fix{f_2}{t_2}$ we get: \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi}} {\Phi\defsubst{t_1\subst{f_1}{\later \fix{f_1}{t_1}}}{t_2\subst{f_2}{\later \fix{f_2}{t_2}}}}$. \\ Since $t\subst{f}{\later \fix{f}{t}} \equiv \fix{f}{t}$, by \rname{CONV}: \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi}}{\Phi\defsubst{\fix{f_1}{t_1}}{\fix{f_2}{t_2}}}$. \\ and since $\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi} \Leftrightarrow \later \Phi\defsubst{\fix{f_2}{t_2}}{\fix{f_2}{t_2}}$, \\ $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later\Phi\defsubst{\fix{f_1}{t_1}}{fix{f_2}{t_2}}}{\Phi\defsubst{fix{f_1}{t_1}}{fix{f_2}{t_2}}}$, \\ and finally, by \rname{L\"ob} we get our goal. \\ \noindent {\bf Case.} $\small\inferrule*[Right=\sf Cons] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{x_2}{A_2}{\Phi_h} \\ \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{xs_2}{\later{\Str{A_2}}}{\Phi_t} \\ \jhol{\Gamma}{\Psi}{\forall x_1,x_2, xs_1, xs_2. \Phi_h\defsubst{x_1}{x_2} \Rightarrow \Phi_t\defsubst{xs_1}{xs_2} \Rightarrow \Phi\defsubst{\cons{x_1}{xs_1}}{\cons{x_2}{xs_2}}}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{xs_1}}{\Str{A_1}}{\cons{x_2}{xs_2}}{\Str{A_2}}{\Phi}}$ \\ \\ Apply the I.H., \rname{$\forall_E$} and \rname{$\To_E$}. \\ \noindent {\bf Case.}$\small\inferrule*[Right=\sf Head] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{\Str{A_1}}{\Phi\defsubst{hd\ \res\ltag}{hd\ \res\rtag}}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{hd\ t_2}{A_2}{\Phi}}$ \\ \\ Trivial by I.H. \\ \noindent {\bf Case.} $\small\inferrule*[Right=\sf Tail] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{\Str{A_2}}{\Phi\defsubst{tl\ \res\ltag}{tl\ \res\rtag}}} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{Str_{A_1}}}{tl\ t_2}{\later{Str_{A_2}}}{\Phi}}$ \\ \\ Trivial by I.H. \\ \noindent {\bf Case.} $\small\inferrule*[Right=\sf Equiv] {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t'_1}{A_1}{t'_2}{A_2}{\Phi} \\ t_1 \equiv t_1' \\ t_2 \equiv t_2' \\ \Delta \mid \Gamma \vdash t_1 : A_1 \\ \Delta \mid \Gamma \vdash t_2 : A_2} {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi}}$ \\ \\ Trivial by I.H. and \rname{Conv}. \\ Most of the proofs for the probabilistic fragment are a consequence of the proof of \autoref{thm:sound-ghol}. The only interesting case is \rname{Markov}. We do the proof directly in RHOL by showing we can derive it from \rname{Fix}. We have the premises: \begin{enumerate} \item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi}$ \item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{h_1}{C_1 \to \Distr(C_1)}{h_2}{C_2 \to \Distr(C_2)}{\psi_3}$ \item $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\psi_4}$ \end{enumerate} where: \begin{align*} \psi_3 &\equiv \forall x_1 x_2. \phi\defsubst{x_1}{x_2} \Rightarrow \diamond_{[ y_1 \ot \res\ltag\ x_1, y_2 \ot \res\rtag\ x_2]}\phi\defsubst{y_1}{y_2} \\ \psi_4 &\equiv \forall x_1\ x_2\ xs_1\ xs_2. \phi\defsubst{x_1}{x_2} \Rightarrow \later[y_1 \ot xs_1, y_2 \ot xs_2]{\Phi} \Rightarrow \Phi\subst{y_1}{\cons{x_1}{xs_1}}\subst{y_2}{\cons{x_2}{xs_2}} \end{align*} If we inline the definition of unfold, we have to prove: \[\begin{aligned}[t] \fix{f}{&\lambda x_1. \lambda h_1. \mlet{z_1}{h_1\ x_1} { \mlet{t_1}{\operatorname{swap_{\later\Distr}^C}(f_1\app \later z_1 \app \later h_1)} { \munit{\cons{x_1}{t_1}}}} }\sim \\ \fix{f}{&\lambda x_2. \lambda h_2. \mlet{z_2}{h_2\ x_2} { \mlet{t_2}{\operatorname{swap_{\later\Distr}^C}(f_2\app \later z_2 \app \later h_2)} { \munit{\cons{x_2}{t_2}}}} }\\\mid\; &\forall x_1 x_2 h_1 h_2. \phi\defsubst{x_1}{x_2} \Rightarrow \psi_3\defsubst{h_1}{h_2} \Rightarrow \diamond_{[y_2 \ot\res\ltag, y_2 \ot\res\rtag]}\Phi \end{aligned} \] We apply \rname{FIX}, \rname{MLET} twice, and then \rname{MUNIT}. The main judgements we have to prove are: \begin{enumerate}[label*=(\alph*)] \item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{h_1\ x_1}{\Distr(C_1)}{h_2\ x_2}{\Distr(C_2)} {\diamond_{[\res\ltag\ot\res\ltag, \res\rtag\ot\res\rtag]}\phi}$ \item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\operatorname{swap_{\later\Distr}^C}(f_1\app \later z_1 \app \later h_1)}{\Distr(\later C_1)} {\operatorname{swap_{\later\Distr}^C}(f_2\app \later z_2 \app \later h_2)}{\Distr(\later C_2)} {\diamond_{[z_1 \ot \res\ltag, z_2 \ot \res\rtag]} \later[y_1 \ot z_1, y_2 \ot z_2]{\Phi}}$ \item $\jgrhol{\Delta}{\Sigma}{\Gamma, x_1,x_2,t_1,t_2}{\Psi, \phi\defsubst{x_1}{x_2}, \later[y_1 \ot t_1, y_2 \ot t_2]{\Phi}}{\cons{y_1}{t_1}}{\Distr(\Str{C_1})}{\cons{y_2}{t_2}}{\Distr(\Str{C_2})} {\Phi\subst{y_1}{\res\ltag}\subst{y_2}{\res\rtag}}$ \end{enumerate} The judgement (a) is a direct consequence of premises (1) and (2), (b) is proven from the inductive hypothesis, and (d) is a direct consequence of (3). This completes the proof. \section{Examples} \subsection{Proof of ZipWith} This example, taken from \cite{BGCMB16}, proves a property about the $\operatorname{ZipWith}$ function, which takes two streams of type A, a function on pairs of elements, and ``zips'' the two streams by applying that function to the elements that are at the same position on the two streams. We want to show that if the function on the elements is commutative, zipping two streams with that function is commutative as well. We can define the zipWith function as: \begin{align*} \operatorname{zipWith} &: (\nat \to \nat \to \nat) \to \Str{\nat} \to \Str{\nat} \to \Str{\nat} \\ \operatorname{zipWith} &\defeq \fix{\operatorname{zipWith}}{\lambda f. \lambda xs. \lambda ys. \cons{(f\ (\hd\ xs)\ (\hd\ ys))}{(\operatorname{zipWith} \app (\tl\ xs) \app (\tl\ ys))}} \end{align*} We prove (omitting types of expressions): $$\jgrholnocnot{\operatorname{zipWith}}{\operatorname{zipWith}}{\Phi}$$ where \begin{align*} \Phi &\defeq \forall f_1 f_2. (f_1 = f_2 \wedge \forall x y. f_1 x y = f_1 y x) \Rightarrow \forall xs_1 xs_2. \forall ys_1 ys_2. (xs_1 = ys_2 \wedge xs_2 = ys_1)\\ & \Rightarrow \res\ltag\ f_1\ xs_1\ ys_1 = \res\rtag\ f_2\ xs_2\ ys_2 \\ A &\defeq (\nat \to \nat \to \nat) \to \Str{\nat} \to \Str{\nat} \to \Str{\nat} \end{align*} The proof proceeds by applying two-sided rules all the way. We invite interested readers to compare this proof with the one given in~\cite{BGCMB16} to see how the two approaches differ. We show how to derive the statement backwards. The derivation begins with the \rname{Fix} rule. Its premise is (omitting constant contexts): \[ \jrhol{\operatorname{zipWith}_1,\operatorname{zipWith}_2:\later A}{\later[\res\ltag \leftarrow \operatorname{zipWith}_1,\res\rtag \leftarrow \operatorname{zipWith}_2] \Phi}{\lambda f_1.(\cdots)}{A} {\lambda f_2.(\cdots)}{A}{\Phi} \] Then we apply the \rname{ABS} rule three times to introduce into the context the logical relations on $f_1,f_2,xs_1,xs_2,ys_1$, and $ys_2$. The premise we need to prove is then: \[ \begin{array}{c} \jrhol{\Gamma}{\Psi}{\cons{(f_1\ (hd\ xs_1)\ (hd\ ys_1))}{(\operatorname{zipWith}_1 \app (tl\ xs_1) \app (tl\ ys_1))}}{\Str{\nat}} {\\ \cons{(f_2\ (hd\ xs_2)\ (hd\ ys_2))}{(\operatorname{zipWith}_2 \app (tl\ xs_2) \app (tl\ ys_2))}}{\Str{\nat}}{\res\ltag = \res\rtag} \end{array} \] where \begin{align*} \Gamma &\defeq \operatorname{zipWith}_1,\operatorname{zipWith}_2 :\later A; f_1,f_2:(\nat\to\nat\to\nat); xs_1,xs_2,ys_1,ys_2: \Str{\nat} \\ \Psi &\defeq \later[\res\ltag \leftarrow \operatorname{zipWith}_1,\res\rtag \leftarrow \operatorname{zipWith}_2] \Phi, (f_1 = f_2 \wedge \forall x y. f x y = f y x), xs_1 = ys_2, xs_2 = ys_1 \end{align*} Now we can apply the \rname{Cons} rule, which has three premises: \begin{enumerate} \item $\jrhol{\Gamma}{\Psi}{f_1\ (hd\ xs_1)\ (hd\ ys_1)}{\nat} {f_2\ (hd\ xs_2)\ (hd\ ys_2)}{\nat}{\res\ltag = \res\rtag}$ \item $\jrhol{\Gamma}{\Psi}{\operatorname{zipWith}_1 \app (tl\ xs_1) \app (tl\ ys_1)}{\later \Str{\nat}} {\operatorname{zipWith}_2 \app (tl\ xs_2) \app (tl\ ys_2)}{\later \Str{\nat}}{\res\ltag = \res\rtag}$ \item $\jhol{\Gamma}{\Psi}{\forall x y xs ys. x=y \Rightarrow xs = ys \Rightarrow \cons{x}{xs} = \cons{y}{ys}}$ \end{enumerate} Premise (3) is easily provable in HOL. To prove premise (1) we first apply the \rname{App} rule twice, and we have to prove the judgments: \begin{itemize} \item $\jrhol{\Gamma}{\Psi}{f_1}{\nat\to\nat\to\nat} {f_2}{\nat\to\nat\to\nat}{ \forall vs_1 vs_2 ws_1 ws_2. \\ vs_1 = hd\ ys_2 \wedge vs_2 = hd\ ys_1 \Rightarrow ws_1 = hd\ xs_2 \wedge ws_2 = hd\ xs_1 \Rightarrow \res\ltag\ vs_1\ ws_1 = \res\rtag\ vs_2\ ws_2}$ \item $\jrhol{\Gamma}{\Psi}{hd\ xs_1}{\nat} {hd\ xs_2}{\nat}{\res\ltag = hd\ ys_2 \wedge \res\rtag = hd\ ys_1}$ \item $\jrhol{\Gamma}{\Psi}{hd\ ys_1}{\nat} {hd\ ys_2}{\nat}{\res\ltag = hd\ xs_2 \wedge \res\rtag = hd\ xs_1}$ \end{itemize} The three can be proven in HOL from the conditions imposed on $f_1,f_2$ and the equalities $xs_1 = ys_2$, $xs_2 = ys_1$. All that remains to prove is premise (2) of the \rname{Cons} application, which, by expanding the definition of $\app$ and using the equational theory of delayed substitutions, can be desugared to: \[\jrhol{\Gamma}{\Psi}{\latern[\xi_1]{(g_1\ t_1\ u_1)}}{\later \Str{\nat}} {\latern[\xi_2]{(g_2\ t_2\ u_2)}}{\later \Str{\nat}} {\latern[\xi_1, \xi_2, [\res\ltag\leftarrow\res\ltag, \res\rtag \leftarrow \res\rtag]]{(\res\ltag = \res\rtag)}}\] where, for $i=1,2$: \[ \xi_i = [g_i \leftarrow \operatorname{zipWith}_i, t_i\leftarrow(tl\ xs_i), u_i \leftarrow (tl\ ys_i)]\] We apply the \rname{Next} rule, and we have the four following premises: \begin{itemize} \item $\jrhol{\Gamma}{\Psi}{\operatorname{zipWith}_1}{\later A} {\operatorname{zipWith}_2}{\later A} {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = \res\rtag \wedge \forall x y. \res\ltag x y = \res\ltag y x)}$ \item $\jrhol{\Gamma}{\Psi}{tl\ xs_1}{\later \Str{\nat}} {tl\ xs_2}{\later \Str{\nat}} {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = tl\ ys_2 \wedge \res\rtag = tl\ ys_1)}$ \item $\jrhol{\Gamma}{\Psi}{tl\ ys_1}{\later \Str{\nat}} {tl\ ys_2}{\later \Str{\nat}} {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = tl\ xs_2 \wedge \res\rtag = tl\ xs_1)}$ \item $\jrhol{\Gamma; g_1,g_2: A; t_1,t_2,u_1,u_2: \Str{\nat}} {\Psi, g_1 = g_2 \wedge \forall x y. g_1 x y = g_1 y x, t_1 = tl\ ys_2 \wedge t_2 = tl\ ys_1, \\ u_1 = tl\ xs_2 \wedge u_2 = tl\ xs_1} {g_1\ t_1\ u_1}{\Str{\nat}} {g_2\ t_2\ u_2}{\Str{\nat}} {\res\ltag = \res\rtag}$ \end{itemize} To prove the first premise we instantiate the inductive hypothesis we got from \rname{Fix}. To prove the second and the third premises we use the equalities $xs_1 = ys_2$, $xs_2 = xs_1$. Finally, the fourth premise is a simple derivation in HOL that follows from the same equalities plus the refinements of $g_1,g_2,t_1,t_2,u_1,u_2$. This concludes the proof. \subsection{Proof of approximation series} We now continue with another example that, while still being fully synchronous (i.e., uses only two-sided rules), goes beyond reasoning about equality of streams, and showcases the flexibility of streams to represent different kinds of information and structures. For instance, streams can be used to represent series of numbers. In this example, we illustrate an instance of a property about series that can be proven in our system. Consider the series $x_0, x_1, \ldots$ for any $p \geq \frac{1}{2}$ and any $a \geq 0$, where $x_0$ is given and: \[ x_{i+1} = px_i + (1-p) \frac{a}{x_i} \] It can be easily shown that if $x_0 \geq \sqrt{a}$, then this series converges \emph{monotonically} from the top to $\sqrt{a}$. In particular, $\displaystyle\lim_{i \to \infty} x_i = \sqrt{a}$. (For $p = \frac{1}{2}$, this is the standard Newton-Raphson series for square-root computation \cite{Scott11}) The interesting relational property is that for smaller $p$, this series converges faster. Concretely, define $f(p, a, x_0, i)$ as the $i$th element of the above series (for the given $p$, $a$ and $x_0$). Then, the relational property to prove is that: \[ \forall p_1\, p_2\, a\, x_0\, i.\, (\dfrac{1}{2}\leq p_1 \leq p_2 \conj x_0 \geq \sqrt{a}) \Rightarrow |f(p_1,a,x_0,i) - \sqrt{a}| \leq |f(p_2,a,x_0,i)- \sqrt{a}|\] We outline the proof of this property. First, note that because convergence is from the top, $|f(p_2,a,x_0,i) - \sqrt{a}| = f(p_2,a,x_0,i) - \sqrt{a}$. Therefore, the property above is the same as: \[ \forall p_1\, p_2\, a\, x_0\, i.\, (\dfrac{1}{2}\leq p_1 \leq p_2\conj x_0 \geq \sqrt{a}) \Rightarrow f(p_2,a,x_0,i) - f(p_1,a,x_0,i) \geq 0\] This is easy to establish by induction on $i$. (Note the importance of the assumption $p \geq \frac{1}{2}$: Without this assumption, convergence is not monotonic, and this relational property may not hold. If we start with $x_0 \leq \sqrt{a}$ instead of $x_0 \geq \sqrt{a}$, we need $p \leq \frac{1}{2}$ for convergence to be monotonic, this time from below.) Now we see how we can encode and prove this as a relational property of a pair of streams. We can define a stream whose elements are the elements of one of this series: \begin{align*} \operatorname{approx\_sqrt} &: \real \to \real \to \real \to \Str{\real} \\ \operatorname{approx\_sqrt} &\defeq \fix{f}{\lambda p. \lambda a. \lambda x. \cons{x}{(f \app \later p \app \later a \app \later (p*x + (1-p)*a/x))}} \end{align*} We prove: \[\jgrholnoc{\operatorname{approx\_sqrt_1}}{\real\to\real\to\real\to \Str{\real}}{\operatorname{approx\_sqrt_2}}{\real\to\real\to\real\to \Str{\real}} {\Phi} \] where \begin{align*} \Phi &\defeq \forall p_1 p_2. (\dfrac{1}{2}\leq p_1 \leq p_2) \Rightarrow \forall a_1 a_2. 0 \leq a_1 = a_2 \Rightarrow \forall x_1 x_2. (0 \leq x_1 \leq x_2 \wedge a_1\leq x_1*x_1) \\ &\Rightarrow \All(\res\ltag\ p_1\ a_1\ x_1, \res\rtag\ p_2\ a_2\ x_2, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a_1 \leq n_1 * n_1) \end{align*} and $\All$ is defined axiomatically as follows: \[\forall s_1,s_2,n_1,n_2. \phi n_1 n_2 \Rightarrow \later[s'_1\ot s_1, s'_2 \ot s_2]{\All(s'_1,s'_2,\lambda x_1 x_2. \phi)} \Rightarrow \All(\cons{n_1}{s_1}, \cons{n_2}{s_2}, \lambda x_1 x_2. \phi)\] The meaning of the judgement is that, if we have two approximation series for the square root of $a$ (formally, we write $a = a_1 = a_2$), with initial guesses $x_1 \leq x_2$, and parameters $1/2 \leq p_1 \leq p_2$, then, at every position, the first series is going to be closer to the root than the second one. Note that we have removed the square roots in the specification by squaring. Let $A =\defeq \real\to\real\to\real\to \Str{\real}$. We will show how to derive the judgment backwards. The proof starts by applying \rname{Fix} which has the premise (omitting constant contexts): \[\jrhol{f_1,f_2 : \later A}{\later[\res\ltag, \res\rtag \leftarrow f_1, f_2]\Phi}{\lambda p_1. \lambda a_1. \lambda x_1. \dots}{A}{\lambda p_2. \lambda a_2. \lambda x_2. \dots}{A} {\Phi} \] and after applying \rname{Abs} three times: \[\begin{array}{c} \jrhol{f_1,f_2 : \later A; p_1,p_2,a_1,a_2,x_1,x_2 : \real} {\\ \later[\res\ltag, \res\rtag \leftarrow f_1, f_2]\Phi, (\dfrac{1}{2}\leq p_1 \leq p_2) , 0 \leq a_1 = a_2 , 0 \leq x_1 \leq x_2, a_1 \leq x_1*x_1} {\\ (\lambda y_1.\cons{x_1}{(f_1\app \later p_1\app \later a_1\app \later y_1)})(p_1*x_1 + (1-p_1)*a_1/x_1)}{\Str{\real}} {\\ (\lambda y_2.\cons{x_2}{(f_2\app \later p_2\app \later a_2\app \later y_2)})(p_2*x_2 + (1-p_2)*a_2/x_2)}{\Str{\real}} {\\ All(\res\ltag, \res\rtag, \lambda n_1 n_2. n_1 \leq n_2)} \end{array} \] Let $\Gamma$ and $\Psi$ denote the typing and logical contexts in the previous judgement. Now we apply \rname{App}, which has two premises: \begin{itemize} \item $\jrhol{\Gamma}{\Psi} {\lambda y_1.\cons{x_1}{(f_1\app \later p_1\app \later a_1\app \later y_1)}}{\real\to \Str{\real}} {\lambda y_2.\cons{x_2}{(f_2\app \later p_2\app\later a_2\app\later y_2)}}{\real\to \Str{\real}} {\\ \forall y_1, y_2. (0 \leq y_1 \leq y_2 \wedge a_1 \leq y_1 * y_1) \Rightarrow All(\res\ltag\ y_1, \res\rtag\ y_2, \lambda n_1 n_2. 0\leq n_1\leq n_2 \wedge a_1 \leq n_1 * n_1)}$ \item $\jrhol{\Gamma}{\Psi} {p_1*x_1 + (1-p_1)*a_1/x_1}{\Str{\real}} {p_2*x_2 + (1-p_2)*a_2/x_2}{\Str{\real}} {\\ 0 \leq \res\ltag \leq \res\rtag \wedge a_1 \leq \res\ltag * \res\ltag}$ \end{itemize} The second premise can be established in Guarded HOL as an arithmetic property in our theory of reals. To prove the first one, we start by applying the \rname{Abs} rule, followed by the \rname{Cons} rule, which has three premises: \begin{enumerate} \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {x_1}{\real} {x_2}{\real} {0 \leq \res\ltag \leq \res\rtag \wedge a \leq \res\ltag * \res\ltag}$ \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {(f_1\app\later p_1\app\later a_1\app\later y_1)}{\later \Str{\real}} {\\ (f_2\app\later p_2\app\later a_2\app\later y_2)}{\later \Str{\real}} {\later[\res\ltag \ot \res\ltag, \res\rtag\ot\res\rtag]All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a \leq n_1 * n_1)}$ \item $\jhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {\forall h_1 h_2 t_1 t_2. 0 \leq h_1 \leq h_2 \Rightarrow a_1 \leq h_1*h_1 \Rightarrow\\ \later[\res\ltag\ot t_1, \res\rtag \ot t_2] All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0\leq n_1 \leq n_2 \wedge a \leq n_1*n_1) \Rightarrow \\ All(\cons{h_1}{t_1}, \cons{h_2}{t_2}, \lambda n_1 n_2. 0\leq n_1 \leq n_2 \wedge a \leq n_1*n_1))}$ \end{enumerate} Premise (1) is just the refinement on $x_1,x_2$, while premise (3) is the axiomatization of $All$. To prove premise (2) one instantiates the induction hypothesis given by the \rname{Fix} rule. In order to do so, we first rewrite the two terms we are comparing to their desugared form: \[\later[f_1' \ot f_1, p_1' \ot \later p_1, a_1' \ot \later a_1, y_1' \ot \later y_1] f_1'\ p_1'\ a_1'\ y_1'\] and \[\later[f_2' \ot f_2, p_2' \ot \later p_2, a_2' \ot\later a_2, y_2' \ot\later y_2] f_2'\ p_2'\ a_2'\ y_2'\] We can also add by \rname{SUB} the same substitutions to the $\later$ in the conclusion, since the substituted variables do not appear in the formula. Then we can apply the \rname{Next} rule, which has the premises: \begin{itemize} \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {f_1}{\later A} {f_2}{\later A} {\later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag]\Phi}$ \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {\later p_1}{\real} {\later p_2}{\real} {\\ \later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] \dfrac{1}{2}\leq \res\ltag \leq \res\rtag}$ \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {\later a_1}{\real} {\later a_2}{\real} {\later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] 0 \leq \res\ltag = \res\rtag}$ \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)} {\later y_1}{\real} {\later y_2}{\real} {\\ \later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] 0 \leq \res\ltag \leq \res\rtag \wedge a_1' \leq \res\ltag * \res\ltag}$ \item $\jrhol{\Gamma, y_1,y_2,p_1',p_2',a_1',a_2',y_1',y_2' : \real; f_1', f_2' : A} {\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1), \Phi\defsubst{f_1'}{f_2'},\\ \dfrac{1}{2}\leq p_1' \leq p_2', 0 \leq a_1' = a_2', 0 \leq y_1' \leq y_2' \wedge a_1' y_1' * y_1'} {\\ f_1'\ p_1'\ a_1'\ y_1'}{\Str{\real}} {f_2'\ p_2'\ a_2'\ y_2'}{\Str{\real}}{All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a \leq n_1 * n_1)}$ \end{itemize} The first four can be proven simply by instantiating and then delaying one of the axioms. The last one is proven by applying \rname{App} three times. This concludes the proof. \subsection{Proof of Cassini's identity} We continue building on the idea from the previous example of using streams to represent series of numbers. This time, we prove a classical identity of the Fibonacci sequence. Since the example requires to observe the stream at different times, we will also have to deal with some asynchronicity on the delayed substitutions. Let $F_n$ be the $n$th Fibonacci number. Cassini's identity states that $F_{n-1} \cdot F_{n+1} - F_{n}^2 = (-1)^{n}$. Cassini's identity can be stated as a stream problem as follows. First, let $F$ be the Fibonnaci stream ($1,1,2,3,5,\ldots$) and $A$ be the stream $1,-1,1,-1,\ldots$ Let $\oplus$ and $\otimes$ be infix functions that add and multiply two streams pointwise. Cassini's identity can then be informally written as: \[ F \otimes \tl(\tl\; F) = \tl(F \otimes F) \oplus A \] In order to formalize Cassini's identity in our system, we first define: \begin{align*} \oplus &: \Str{\nat} \to \Str{\nat} \to \Str{\nat} &\otimes &: \Str{\nat} \to \Str{\nat} \to \Str{\nat} \\ \oplus &\defeq \begin{array}{l}\fix{f}{\lambda s.\lambda t\\ \cons{(\hd\ x + \hd\ y)}{(f \app (\tl\ x) \app (\tl\ y))}}\end{array} \quad \quad &\otimes &\defeq \begin{array}{l}\fix{f}{\lambda s.\lambda t\\ \cons{(\hd\ x * \hd\ y)}{(f \app (\tl\ x) \app (\tl\ y))}}\end{array} \end{align*} Then we define $F$ and $A$ as the fixpoints of the equations: \begin{align*} F &\defeq \fix{F}{\cons{1}{\later[ F' \ot F] (\cons{1}{\later [ T \ot \tl\ F' ] ( F' \oplus T)})}} \\ A &\defeq \fix{A}{\cons{1}{\later(\cons{-1}{A})}} \end{align*} We prove (using prefix notation for $\oplus$ and $\otimes$): \[\jgrholnoc {\later [ T_1 \ot \tl\ F] \otimes \app (\later F) \app \tl\ T_1 }{\later\later \Str{\nat}} {\oplus \app \tl(F \otimes F) \app (\later A)}{\later \Str{\nat}} { \res\ltag = \later \res\rtag} \] The proof combines applications of two-sided rules and one-sided rules; in particular, we use the rule \rname{NEXT-L} to proceed with the proof for a judgement where the left expression is delayed twice and the right expression is delayed once. By conversion, in the logic we can prove the following equalities: \[ \Psi \defeq \left\{\begin{aligned} F &= \cons{1}{\later(\cons{1}{\later [ T \ot tl\ F ] ( F \oplus T)})}, \\ A &= \cons{1}{\later(\cons{-1}{\later A})} \end{aligned}\right\} \] Using these equalities, and desugaring the applications, the judgment we want to prove is (omitting constant contexts): \[\begin{array}{c}\jrhol{F,A: \Str{\nat}}{\Psi} {\later [ T_1 \ot tl\ F] \later [T_1' \ot tl\ T_1] (F \otimes T_1') }{\later\later \Str{\nat}} {\\ \later [T_2 \ot tl(F \otimes F)] (T_2 \oplus A)}{\later \Str{\nat}} {\\ \later[\res\ltag' \ot \res\ltag, T_1 \ot tl\ F] \later [\res\ltag''\ot \res\ltag, \res\rtag'\ot\res\rtag, T_1' \ot tl\ T_1, T_2' \ot tl(F \otimes F)] \res\ltag'' = \res\rtag'} \end{array} \] Notice that on the left, since we want to apply tail twice to $F$, we need to delay the term twice so that $F$ and $tl\ tl\ F$ have the same type. On the right, we just need to delay the term once. As for the logical conclusion, $\res\ltag$ needs to be delayed twice, while $\res\rtag$ only once. The way to do this is by having $\res\ltag$ appear on the two substitutions but $\res\rtag$ only on the inner one. We start by applying \rname{NEXT-L}, which has the two following premises: \begin{itemize} \item $\juhol{F,A: \Str{\nat}}{\Psi}{tl\ F }{\later \Str{\nat}}{\later[\res' \ot \res] tl\ F = \later \res'}$ \item $\jrhol{F,A, T_1: \Str{\nat}}{\Psi, tl\ F = \later T_1}{\later [T_1' \ot tl\ T_1] (F \otimes T_1') }{\later \Str{\nat}} {\\ \later [T_2 \ot tl(F \otimes F)] (T_2 \oplus A)}{\later \Str{\nat}} {\later [\res\ltag''\ot \res\ltag, \res\rtag'\ot\res\rtag, T_1' \ot tl\ T_1, T_2' \ot tl(F \otimes F)] \res\ltag'' = \res\rtag'}$ \end{itemize} The first premise is trivial. We continue by applying \rname{NEXT} to the second, which has the following premises: \begin{itemize} \item $\jrhol{F,A, T_1: \Str{\nat}}{\Psi, tl\ F = \later T_1}{tl\ T_1}{\later \Str{\nat}}{tl(F \otimes F)}{\later \Str{\nat}} {\\ \later[\res\ltag' \ot \res\ltag, \res\rtag' \ot \res\rtag] T_1 = \later \res\ltag' \wedge \res\ltag'\otimes\res\ltag' = \res\rtag'}$ \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{F \otimes T_1'}{\Str{\nat}}{T_2 \oplus A}{\Str{\nat}}{\res\ltag = \res\rtag}$ \end{itemize} Again, the first premise is trivial. We apply \rname{APP} twice to the second, and we have to prove: \begin{itemize} \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{F}{\Str{\nat}}{A}{\Str{\nat}}{\\ \res\ltag = F \wedge \res\rtag = A}$ \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{T_1'}{\Str{\nat}}{T_2}{\Str{\nat}} {\\ F = \cons{1}{\later(\cons{1}{\later T_1})} \wedge \res\ltag \otimes \res\rtag = \res\rtag}$ \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2} {\\ \otimes}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\oplus}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}} {\forall X_1 X_2 Y_1 Y_2. X_1 = F \wedge X_2 = A \Rightarrow F = \cons{1}{\later(\cons{1}{\later Y_1})} \wedge Y_1 \otimes Y_1 = Y_2 \Rightarrow \res\ltag\ X_1\ Y_1 = \res\rtag\ X_2\ Y_2}$ \end{itemize} The two first premises are easy to prove. We will show how to prove the last one. For this, we need a stronger induction hypothesis for $\hat{\oplus}$ and $\hat{\otimes}$. We propose the following: \[\begin{array}{c} \forall g_1,g_2,b_1,G,B. G = \conshat{g_1}{\conshat{g_2}{(G\hat{\oplus} (\hat{tl} G))}} \wedge b_1 = g_1^2 + g_1 g_2 - g_2^2 \wedge B = \conshat{b_1}{\conshat{-b_1}{B}} \\ \Rightarrow G \hat{\otimes} \hat{tl}(\hat{tl} G) = \hat{tl} (G \hat{\otimes} G) \hat{\oplus} B \end{array} \] We then use the \rname{SUB} rule to strengthen the inductive hypothesis, and now the new judgement to prove is: \[\begin{array}{c} \jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2} {\\ \otimes}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\oplus}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}} {\\ \forall X_1 X_2 Y_1 Y_2. (\exists g_1, g_2, b_1, G, B. G = \cons{g_1}{\later(\cons{g_2}{\later [ G' \ot tl\ G ] ( G \oplus G')})} \wedge b_1 = g_1^2 + g_1 g_2 - g_2^2 \wedge \\ B = \cons{b_1}{\later(\cons{-b_1}{\later B})} \wedge X_1 = G \wedge X_2 = B \wedge X_1 = \cons{1}{\later(\cons{1}{\later Y_1})} \wedge Y_1 \otimes Y_1 = Y_2) \Rightarrow \res\ltag\ X_1\ Y_1 = \res\rtag\ X_2\ Y_2} \end{array} \] Let $\Gamma'$, $\Psi'$ and $\Phi_{IH}$ denote respectively the typing context, logical context and logical conclusion of the previous judgement. The premise of the FIX rule is: \[\begin{array}{c} \jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat})}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}} {\\ \fix{f_1}{\lambda X_1. \lambda Y_1. \dots}}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\fix{f_2}{\lambda X_2. \lambda Y_2. \dots}}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\Phi_{IH}} \end{array} \] Let $\Phi_E$ denote the existential clause in $\Phi_{IH}$. After applying \rname{ABS} twice, we have: \[\begin{array}{c} \jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E} {\\ \cons{(hd\ X_1)*(hd\ Y_1)}{f_1 \app (tl\ X_1) \app (tl\ Y_1)}}{\Str{\nat}}{\cons{(hd\ X_2)+(hd\ Y_2)}{f_2 \app (tl\ X_2) \app (tl\ Y_2)}}{\Str{\nat}}{\res\ltag = \res\rtag} \end{array}\] And then we apply \rname{Cons} to prove equality on the heads and the tails: \begin{itemize} \item $\jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E} {\\ (hd\ X_1)*(hd\ Y_1)}{\nat}{(hd\ X_2)+(hd\ Y_2)}{\nat}{\res\ltag = \res\rtag}$ \item $\jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E} {\\ f_1 \app (tl\ X_1) \app (tl\ Y_1)}{\later \Str{\nat}}{f_2 \app (tl\ X_2) \app (tl\ Y_2)}{\later \Str{\nat}}{\res\ltag = \res\rtag}$ \end{itemize} To prove the first one we notice that $hd X_1 * hd Y_1 = g_1 * (g_1 + g_2) = g_1^2 + g_2*g_1 = g_2^2 + g_1^2 + g_1 * g_2 - g_2^2 = hd X_2 * hd Y_2$. To prove the second one we need to check that $tl X_1,tl Y_1, tl X_2,tl Y_2$ satisfy the precondition of the inductive hypothesis. In particular, we need to check that \[-b_1 = -g_1^2 - g_1 g_2 + g_2^2 = g_2^2 + g_2 (g_1 + g_2) - (g_1 + g_2)^2\] which is can be proven by arithmetic computation. \section{Unary fragment} In this section we introduce a unary system to prove properties about a single term of the guarded lambda calculus. We will start by adding some definitions Guarded HOL for the unary diamond monad, following by the derivation rules for both the non-probabilistic and the probabilistic system, plus the metatheory and an example. \subsection{Unary fragment of GHOL} The unary semantics of the diamond monad are: \[\sem{\diamond_{[x\leftarrow t]}\phi}_i\defeq \left\{(\delta,\gamma) \isetsep \operatorname{Pr}_{v\ot \left(\sem{t}_i(\delta,\gamma)\right)} [(\delta, (\gamma, v))\in \sem{\phi}_i] = 1 \right\} \] The rules are on Figure~\ref{fig:u-prob-ghol} \begin{figure*}[!htb] \small \begin{mathpar} \infer[\sf MONO1] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\gets t]}\psi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\gets t]}\phi} \and \jghol{\Delta}{\Sigma}{\Gamma, x : C}{\Psi, \phi}{\psi} } \and \infer[\sf UNIT1]{\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[x\leftarrow \munit{t}]} \phi}} {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \phi[t/x]}} \and \infer[\sf MLET1] {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[y\leftarrow \mlet{x}{t}{t'}]} \psi}}{ \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\leftarrow t]} \phi} \and \jghol{\Delta}{\Sigma}{\Gamma,x:C}{\Psi,\phi}{ \diamond_{[y\leftarrow t']}\psi}} \end{mathpar} \caption{Rules for the unary diamond modality}\label{fig:u-prob-ghol} \end{figure*} \subsection{Guarded UHOL} We start by defining the Guarded UHOL system, which allows us to prove logical properties of a term of the Guarded Lambda Calculus. More concretely, judgements have the form: \[ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi} \] where $t$ is a term well-typed in the dual context $\Delta \mid \Gamma$ and $\phi$ is a logical formula well-typed in the context $\Delta \mid \Gamma, \res : \sigma$ and that can refer to $t$ via the special variable $\res$. The logical contexts $\Sigma$ and $\Psi$ consist respectively of refinements over the contexts $\Delta$ and $\Gamma$. \subsection{Derivation rules} The rule \rname{Next} corresponds to the introduction of the later modality. A refinement $\Phi_i$ is proven on every term in the substitution, and using those as a premise, a refinement $\Phi$ is proven on $t$. In the notation $\later[\res \ot \res]\Phi$ the first $\res$ is the variable bound by the delayed substitution inside $\Phi$ while the second $\res$ is the distinguished variable in the refinement that refers to the term that is being typed. In other words, $t$ satisfies $\later[\res \ot \res]\Phi$ if $\later[\res \ot t]\Phi$. The rule \rname{Prev} corresponds to the elimination of the later modality. If we can prove $\later \phi$ in a constant context, then we can also prove $\phi$. The rule \rname{Box} applies the constant modality on a formula that can be proven on a constant context. The rule \rname{LetBox} removes the constant modality from a formula $\Phi$ by using it as a constant premise to prove another formula $\Phi'$. The rule \rname{LetConst} shifts constant terms between contexts. The rule \rname{Fix} introduces a fixpoint and proves a refinement on it by Loeb induction. The rule \rname{Cons} proves a property on a stream from a refinement on its head and its tail. The rule \rname{ConsHat} is the analogue of \rname{Cons} to build constant streams. In particular, the $\hat{::}$ operator can be defined as $\lambda x. \lambda s. \letbox{(y,t)}{(x,s)}{\boxx(\cons{y}{\later{t}})}$. Conversely the rules \rname{Head} and \rname{Tail} respectively prove a property on the head and the tail of a stream from a property on the full stream. \begin{figure*}[!htb] \small \begin{mathpar} \inferrule*[Right=\sf Next] {\jguhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi, \Phi_1\subst{\res}{x_1}, \dots, \Phi_n\subst{\res}{x_n}}{t}{A}{\Phi} \\ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\later{A_1}}{\dsubst{\res}{\res}{\Phi_1}} \\ \dots \\ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_n}{\later{A_n}}{\dsubst{\res}{\res}{\Phi_n}}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\nextt{x_1 \leftarrow t_1, \dots, x_n \leftarrow t_n}{t}}{\later{A}}{\dsubst{x_1,\dots,x_n,\res}{t_1,\dots,t_n,\res}{\Phi}}} \inferrule*[Right=\sf Prev] {\jguhol{\Delta}{\Sigma}{\cdot}{\cdot}{t}{\later A}{\dsubst{\res}{\res}{\Phi}}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t}}{A}{\Phi}} \inferrule*[Right=\sf Box] {\jguhol{\Delta}{\Sigma}{\cdot}{\cdot}{t}{A}{\Phi}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t}}{\square A}{\square \Phi\subst{\res}{\letbox{x}{\res}{x}}}} \inferrule*[Right=\sf LetBox] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{\square B}{\square \Phi\subst{\res}{\letbox{x}{\res}{x}}} \\ \jguhol{\Delta, x : B}{\Sigma, \Phi\subst{\res}{x}}{\Gamma}{\Psi}{t}{A}{\Phi'}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x}{u}{t}}{A}{\Phi'}} \inferrule*[Right=\sf LetConst] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{B}{\Phi} \\ \jguhol{\Delta, x : B}{\Sigma, \Phi\subst{\res}{x}}{\Gamma}{\Psi}{t}{A}{\Phi'} \\ B,\Phi\ \text{constant} \\ FV(\Phi)\cap FV(\Gamma) = \emptyset} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x}{u}{t}}{A}{\Phi'}} \inferrule*[Right=\sf Fix] {\jguhol{\Delta}{\Sigma}{\Gamma, f:\later{A}}{\dsubst{\res}{f}{\Phi}}{t}{A}{\Phi}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{fix f. t}{A}{\Phi}} \inferrule*[Right=\sf Cons] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x}{A}{\Phi_h} \\ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs}{\later{\Str{A}}}{\Phi_t} \\ \jhol{\Gamma}{\Psi}{\forall x,xs. \Phi_h\subst{\res}{x} \Rightarrow \Phi_t\subst{\res}{xs} \Rightarrow \Phi\subst{\res}{\cons{x}{xs}}}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x}{xs}}{\Str{A}}{\Phi}} \inferrule*[Right=\sf ConsHat] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x}{A}{\Phi_h} \\ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs}{\square \Str{A}}{\square \Phi_t} \\ \jhol{\Gamma}{\Psi}{\forall x,xs. \Phi_h\subst{\res}{x} \Rightarrow \Phi_t\subst{\res}{xs} \Rightarrow \Phi\subst{\res}{\conshat{x}{xs}}} \\ A,\Phi_h \text{ constant} } {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\conshat{x}{xs}}{\square \Str{A}}{\square \Phi}} \inferrule*[Right=\sf Head] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Str{A}}{\Phi\subst{\res}{hd\ \res}}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t}{A}{\Phi}} \inferrule*[Right=\sf Tail ] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Str{A}}{\Phi\subst{\res}{tl\ \res}}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t}{\later{Str_A}}{\Phi}} \end{mathpar} \caption{Guarded Unary Higher-Order Logic rules}\label{fig:uhol} \end{figure*} The intended meaning for a judgment $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\tau}{\phi}$ is: ``For every valuations $\delta$, $\gamma$ of $\Delta$ and $\Gamma$, \[\sem{\Delta\mid\Gamma\vdash \square\Sigma}(\delta,\gamma) \wedge \sem{\Delta\mid\Gamma\vdash \Psi}(\delta,\gamma) \To \sem{\Delta\mid\Gamma, \res:\tau\vdash \Sigma}(\delta,\langle \gamma, \sem{\Delta\mid\Gamma\vdash t}(\delta,\gamma)\rangle) \text{''} \] \subsection{Metatheory} We now the most interesting metatheoretical properties of Guarded UHOL. In particular, Guarded UHOL is equivalent to Guarded HOL: \begin{theorem}[Equivalence with Guarded HOL] \label{thm:equiv-uhol-hol} For every contexts $\Delta,\Gamma$, type $\sigma$, term $t$, sets of assertions $\Sigma,\Psi$ and assertion $\phi$, the following are equivalent: \begin{itemize} \item $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi}$ \item $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res}{t}}$ \end{itemize} \end{theorem} The proof is analogous to the relational case The previous result allows us to lift the soundness result from Guarded HOL to Guarded UHOL. \begin{corollary}[Soundness and consistency]\label{cor:uhol:sound} If $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi}$, then for every valuations $\delta \models \Delta$, $\gamma\models\Gamma$: \[\sem{\Delta \vdash \Sigma}(\delta) \wedge \sem{\Delta \mid \Gamma \vdash \Psi}(\delta,\gamma) \Rightarrow \sem{\Delta \mid \Gamma,\res:\sigma \vdash \phi}(\delta, \gamma[\res \ot \sem{\Delta \mid \Gamma \vdash t}(\delta,\gamma)])\] In particular, there is no proof of $\jguhol{\Delta}{\emptyset}{\Gamma}{\emptyset}{t}{\sigma}{\bot}$ in Guarded UHOL. \end{corollary} \subsection{Probabilistic extension} We comment on the rules, starting from the rules of the unary logic. There are three new rules for the probabilistic case, and they all establish that an expression $u$ of type $\Distr(D)$ satisfies the assertion $\diamond_{[y\leftarrow \res]} \phi$, i.e.\, for every element $v$ in the support of (the interpretation of) $u$, the interpretation of $\phi$ with the valuation $[y \mapsto v]$ is true. This intuition is captured by the rule $[\textsf{SUPP}]$, which can be used in particular in case $u$ is a primitive distribution. The rule [\textsf{UNIT}] considers the case where $u$ is of the form $\munit{t}$; in this case, it is clearly sufficient to know that $\phi[t/y]$ is valid. The rule [\textsf{MLET}] simply captures the fact that the support of $\mlet{x}{t}{t'}$ is the disjoint union of the support of $t'$ under all the assignments of $x$ to values in the support of $t$. \begin{figure*}[!htb] \small \textbf{Guarded UHOL} \begin{mathpar} \infer[\sf UNIT] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\munit{t}}{\Distr(C)}{ \diamond_{[y\leftarrow \res]} \Phi}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{C}{\Phi [\res/y]}} \and \infer[\sf MLET] {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x}{t}{t'}}{ \Distr(D)}{\diamond_{[y\leftarrow \res]} \psi}} {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Distr(C)}{\diamond_{[x\leftarrow r]} \Phi} \and \jguhol{\Delta}{\Sigma}{\Gamma,x:C}{\Psi,\phi }{t'}{\Distr(D)}{\diamond_{[y\leftarrow \res]} \psi}} \infer[\sf SUPP]{\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{\Distr(D)}{ \diamond_{[y\leftarrow \res]}\phi}}{\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \Pr_{z\sim u}[\phi[z/y]]=1}} \end{mathpar} \caption{Proof rules for probabilistic constructs -- unary case}\label{fig:puhol} \end{figure*} Finally, we prove an embedding lemma for Guarded UHOL. The proof can be carried by induction on the structure of derivations, or using the equivalence between Guarded UHOL and Guarded HOL (Theorem~\ref{thm:equiv-uhol-hol}). \begin{lem}[Embedding lemma]\label{lem:emb-uhol-rhol} Assume that: \begin{itemize} \item $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{\phi}$ \item $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_2}{\sigma_2}{\phi'}$ \end{itemize} Then $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{ \phi\subst{\res}{\res\ltag}\land \phi'\subst{\res}{\res\rtag}}$. \end{lem} \subsection{Unary example: Every two} We define the $every2$ function, which receives a stream and returns another stream consisting of the elements at even positions in the input stream. Note that this function, while productive, cannot be built with the type $Str \to Str$, since we need to take twice the tail of the argument, which would have type $\later\later Str$, and then a $Str$ cannot be built. Instead, we need to use the constant modality as follows: \begin{align*} every2 &: \square Str \to Str \\ every2 &\defeq \fix{every2}{\lambda s. \cons{\hat{hd}(\hat{tl}\ s)}{(every2 \app {\rm next} (\hat{tl}(\hat{tl}\ s)))}} \end{align*} Where the $\hat{hd}$ and $\hat{tl}$ functions are not the native ones, but rather they are defined as: \begin{align*} \hat{hd} &: \square Str \to \nat &\hat{tl} &: \square Str \to \square Str \\ \hat{hd} &\defeq \lambda s. \letbox{x}{s}{hd\ x} \quad\quad &\hat{tl} &\defeq \lambda s. \letbox{x}{s}{\boxx{(\prev{(tl\ x)})}} \end{align*} The property we want to prove is: \[ \jguhol{\cdot}{\cdot}{ones : \square Str}{\Psi}{every2}{\square Str \to Str}{\forall s. s = ones \Rightarrow \res\ s = (\letbox{x}{s}{x})} \] where $ones$ is the constant stream containing only the number 1 defined as: \[ ones \defeq \boxx{(\fix{f}{\cons{1}{f}})} \] For which we can prove the following properties: \[ \Psi \defeq \hat{hd}\ ones = 1, \hat{tl}\ ones = ones \] In the rest of the proof we omit the empty contexts $\Delta$ and $\Sigma$. We start by applying the \rname{Fix} rule, which has the premise: \[ \begin{array}{c} \juhol{ones : \square Str, every2:\later{(\square Str \to Str)}}{\Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}} {\\ \lambda s. (\cdots)}{\square Str \to Str}{\forall s. s = ones \Rightarrow (\res\ s) = \letbox{x}{s}{x}} \end{array} \] We apply the \rname{Abs} rule inmediately after: \[ \begin{array}{c} \juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones} {\\ \cons{\hat{hd}(\hat{tl}\ s)}{(every2 \app \later (\hat{tl}(\hat{tl}\ s)))}}{Str}{\res = \letbox{x}{s}{x}} \end{array} \] By \rname{SUB} and the equivalence $\letbox{x}{s}{x} \equiv \letbox{x}{ones}{x}$, we can change the conclusion of the judgement. Now we use the \rname{Cons} rule, which has three premises: \begin{enumerate} \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones} { \hat{hd}(\hat{tl}\ s)}{\nat}{\res = 1}$ \\ \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones} {\\(every2 \app \later(\hat{tl}(\hat{tl}\ s)))}{Str}{\dsubst{\res}{\res}{\res = \letbox{x}{ones}{x}}}$ \\ \item $\jhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \dsubst{\res}{every2}{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones} {\\ \forall y, ys. y = 1 \Rightarrow \later[zs \ot ys]{(zs = \letbox{x}{ones}{x})} \Rightarrow \cons{y}{ys} = (\letbox{x}{ones}{x})}$ \end{enumerate} Premises (1) is a consequence of the properties of $ones$. To prove premise (3) we reduce the letbox with the box inside $ones$, and do some reasoning using the definition of the fixpoint. To prove the premise (2) we first desugar the term we are typing: \[every2 \app \later(\hat{tl}(\hat{tl}\ s))) \defeq \later[g \leftarrow every2, t \leftarrow \later(\hat{tl}(\hat{tl}\ s))]{g t} \] and then we apply \rname{Next} which has the following premises: \begin{itemize} \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \dsubst{\res}{every2}{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones} {\\ every2}{\later{(\square Str \to Str)}}{\dsubst{\res}{\res}{(\forall s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}}$ \\ \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str} {\\ \Psi, \dsubst{\res}{every2}{(\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}, s = ones} {\\ \later(\hat{tl}(\hat{tl}\ s))}{\later\square Str}{\dsubst{\res}{\res}{(\res = ones)}}$ \\ \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str, g: \square Str \to Str, t: \square Str} {\\ \Psi, \dsubst{\res}{every2}{(\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}, s = ones, \\ \forall s. s = ones \Rightarrow g\ s = \letbox{x}{s}{x}, t = ones} {g\ t}{Str}{\res = (\letbox{x}{ones}{x})}$ \end{itemize} The first premise is just an application of the \rname{Var} rule. The second premise can be proven as a consequence of the properties of $ones$. Finally, the third premise can be proven with some simple logical reasoning in HOL. This concludes the proof. \end{document}