\documentclass{mscs} \usepackage{comment} \newif\ifshowauxproof \showauxprooffalse \newcommand{\auxprooffont}{\small} \makeatletter \newcommand{\auxproof}[1]{\ifshowauxproof {\auxprooffont \textbf{AuxP} #1 \textbf{AuxQ}}\else\@bsphack\@esphack\fi} \makeatother \ifshowauxproof \newenvironment{Auxproof} {\par\auxprooffont\noindent\textbf{AUX-PROOF}\dotfill\par \noindent\ignorespaces} {\par\noindent\textbf{AUX-QED}\dotfill\par \noindent\ignorespacesafterend} \else \excludecomment{Auxproof} \fi \usepackage[utf8]{inputenc} \usepackage[T1]{fontenc} \usepackage{textcomp} \usepackage{lmodern} \usepackage{exscale} \usepackage{graphicx,color} \usepackage{hyperref} \usepackage{etoolbox} \usepackage{xparse} \usepackage{xspace} \usepackage{latexsym} \usepackage{amssymb,amsmath} \usepackage{mathtools} \usepackage{stmaryrd} \usepackage{bussproofs} \usepackage{nicefrac} \usepackage[noadjust]{cite} \usepackage{enumitem} \usepackage{microtype} \usepackage{wrapfig} \usepackage{listings} \usepackage{fancybox} \usepackage{todonotes} \presetkeys{todonotes}{size=\footnotesize,color=orange!30}{} \newcommand{\Kcomment}[2][]{\todo[color=blue!20,#1]{\textbf{K:} #2}} \newcommand{\Bcomment}[2][]{\todo[color=blue!20,#1]{\textbf{B:} #2}} \allowdisplaybreaks[1] \usepackage{xy} \xyoption{all} \xyoption{tips} \SelectTips{cm}{} \usepackage{tikz} \usetikzlibrary{circuits.ee.IEC} \usetikzlibrary{shapes,shapes.geometric} \newlength\stateheight \setlength\stateheight{0.55cm} \newlength\minimumstatewidth \setlength\minimumstatewidth{0.8cm} \tikzset{width/.initial=\minimummorphismwidth} \tikzset{colour/.initial=white} \newif\ifblack\pgfkeys{/tikz/black/.is if=black} \newif\ifwedge\pgfkeys{/tikz/wedge/.is if=wedge} \newif\ifvflip\pgfkeys{/tikz/vflip/.is if=vflip} \newif\ifhflip\pgfkeys{/tikz/hflip/.is if=hflip} \newif\ifhvflip\pgfkeys{/tikz/hvflip/.is if=hvflip} \pgfdeclarelayer{foreground} \pgfdeclarelayer{background} \pgfsetlayers{background,main,foreground} \def\thickness{0.4pt} \makeatletter \pgfkeys{/tikz/on layer/.code={ \pgfonlayer{#1}\begingroup \aftergroup\endpgfonlayer \aftergroup\endgroup }, /tikz/node on layer/.code={ \gdef\node@@on@layer{\setbox\tikz@tempbox=\hbox\bgroup\pgfonlayer{#1}\unhbox\tikz@tempbox\endpgfonlayer\pgfsetlinewidth{\thickness}\egroup} \aftergroup\node@on@layer }, /tikz/end node on layer/.code={ \endpgfonlayer\endgroup\endgroup } } \def\node@on@layer{\aftergroup\node@@on@layer} \makeatother \makeatletter \pgfdeclareshape{state} { \savedanchor\centerpoint { \pgf@x=0pt \pgf@y=0pt } \anchor{center}{\centerpoint} \anchorborder{\centerpoint} \saveddimen\overallwidth { \pgf@x=3\wd\pgfnodeparttextbox \ifdim\pgf@x<\minimumstatewidth \pgf@x=\minimumstatewidth \fi } \savedanchor{\upperrightcorner} { \pgf@x=.5\wd\pgfnodeparttextbox \pgf@y=.5\ht\pgfnodeparttextbox \advance\pgf@y by -.5\dp\pgfnodeparttextbox } \anchor{A} { \pgf@x=-\overallwidth \divide\pgf@x by 4 \pgf@y=0pt } \anchor{B} { \pgf@x=\overallwidth \divide\pgf@x by 4 \pgf@y=0pt } \anchor{text} { \upperrightcorner \pgf@x=-\pgf@x \ifhflip \pgf@y=-\pgf@y \advance\pgf@y by 0.4\stateheight \else \pgf@y=-\pgf@y \advance\pgf@y by -0.4\stateheight \fi } \backgroundpath { \begin{pgfonlayer}{foreground} \pgfsetstrokecolor{black} \pgfsetlinewidth{\thickness} \pgfpathmoveto{\pgfpoint{-0.5*\overallwidth}{0}} \pgfpathlineto{\pgfpoint{0.5*\overallwidth}{0}} \ifhflip \pgfpathlineto{\pgfpoint{0}{\stateheight}} \else \pgfpathlineto{\pgfpoint{0}{-\stateheight}} \fi \pgfpathclose \ifblack \pgfsetfillcolor{black!50} \pgfusepath{fill,stroke} \else \pgfusepath{stroke} \fi \end{pgfonlayer} } } \makeatother \tikzset{inline text/.style = {text height=1.2ex, text depth=0.25ex,yshift=0.5mm}} \tikzset{arrow box/.style = {rectangle,inline text,fill=white,draw, minimum height=5mm,yshift=-0.5mm,minimum width=5mm}} \tikzset{dot/.style = {inner sep=0mm,minimum width=1mm,minimum height=1mm, draw,shape=circle}} \tikzset{white dot/.style = {dot,fill=white,text depth=-0.2mm}} \tikzset{scalar/.style = {diamond,draw,inner sep=1pt}} \tikzset{copier/.style = {dot,fill,text depth=-0.2mm}} \tikzset{discarder/.style = {circuit ee IEC,ground,rotate=90,scale=0.8,xshift=.55ex}} \tikzset{xshiftu/.style = {shift = {(#1, 0)}}} \tikzset{yshiftu/.style = {shift = {(0, #1)}}} \newenvironment{myproof}[1][Proof ]{ \begin{trivlist}\item[\hskip \labelsep {\bfseries #1}]}{ \end{trivlist}} \newcommand{\pullback}[1][dr]{\save*!/#1-1.2pc/#1:(-1,1)@^{|-}\restore} \makeatother \newdir{ >}{{}*!/-7.5pt/@{>}} \newdir{|>}{!/4.5pt/@{|}*:(1,-.2)@^{>}*:(1,+.2)@_{>}} \newdir{ |>}{{}*!/-3pt/@{|}*!/-7.5pt/:(1,-.2)@^{>}*!/-7.5pt/:(1,+.2)@_{>}} \newcommand{\xyline}[2][]{\ensuremath{\smash{\xymatrix@1#1{#2}}}} \newcommand{\xyinline}[2][]{\ensuremath{\smash{\xymatrix@1#1{#2}}}^{\rule[8.5pt]{0pt}{0pt}}} \newcommand{\filter}{\raisebox{5.5pt}{$\xymatrix@=6pt@H=0pt@M=0pt@W=4pt{\\ \ar@{>->}[u]}$}} \newcommand{\ideal}{\raisebox{1pt}{$\xymatrix@=5pt@H=0pt@M=0pt@W=4pt{\ar@{>->}[d] \\ \mbox{}}$}} \makeatletter \newcommand{\QEDbox}{\square} \newcommand{\QED}{\hspace*{\fill}$\QEDbox$} \newcommand{\after}{\mathrel{\circ}} \newcommand{\klafter}{\mathrel{\bullet}} \newcommand{\scalar}{\mathrel{\bullet}} \newcommand{\cat}[1]{\ensuremath{\mathbf{#1}}} \newcommand{\Cat}[1]{\ensuremath{\mathbf{#1}}} \newcommand{\op}{\ensuremath{^{\mathrm{op}}}} \newcommand{\idmap}[1][]{\ensuremath{\mathrm{id}_{#1}}} \newcommand{\orthogonal}{\mathrel{\bot}} \newcommand{\EfProb}{\textrm{EfProb}\xspace} \newcommand{\Python}{\textrm{Python}\xspace} \newcommand{\BifRel}{\Cat{BifRel}\xspace} \newcommand{\dBisRel}{\Cat{dBisRel}\xspace} \newcommand{\BifMRel}{\Cat{BifMRel}\xspace} \newcommand{\Sets}{\Cat{Sets}\xspace} \newcommand{\Vect}{\Cat{Vect}\xspace} \newcommand{\Conv}{\Cat{Conv}\xspace} \newcommand{\EA}{\Cat{EA}\xspace} \newcommand{\EMod}{\Cat{EMod}\xspace} \newcommand{\AEMod}{\Cat{AEMod}\xspace} \newcommand{\BEMod}{\Cat{BEMod}\xspace} \newcommand{\BCM}{\Cat{BCM}\xspace} \newcommand{\BA}{\Cat{BA}\xspace} \newcommand{\CH}{\Cat{CH}\xspace} \newcommand{\CCH}{\Cat{CCH}\xspace} \newcommand{\CCHobs}{\CCH_{\mathrm{obs}}} \newcommand{\Algobs}{\Alg_{\mathrm{obs}}} \newcommand{\poVectu}{\Cat{poVectu}\xspace} \newcommand{\OUS}{\Cat{OUS}\xspace} \newcommand{\BOUS}{\Cat{BOUS}\xspace} \newcommand{\reals}{\ensuremath{\mathbb{R}}} \newcommand{\Pow}{\mathcal{P}} \newcommand{\nePow}{\mathcal{P}_{\!\!\!\bullet}} \newcommand{\Dst}{\mathcal{D}} \newcommand{\Giry}{\mathcal{G}} \newcommand{\Exp}{\mathcal{E}} \newcommand{\UF}{\ensuremath{\mathcal{U}{\kern-.75ex}\mathcal{F}}} \newcommand{\Mlt}{\mathcal{M}} \newcommand{\Powfin}{\mathcal{P}_{\mathit{fin}}} \newcommand{\Ef}{\ensuremath{\mathrm{Ef}}} \newcommand{\pr}{\ensuremath{\mathrm{Pr}}} \newcommand{\DM}{\ensuremath{\mathrm{DM}}} \newcommand{\tr}{\ensuremath{\mathrm{tr}}\xspace} \newcommand{\ev}{\ensuremath{\mathrm{ev}}\xspace} \newcommand{\gr}{\ensuremath{\mathrm{gr}}\xspace} \newcommand{\cv}{\ensuremath{\mathrm{cv}}\xspace} \newcommand{\ch}{\ensuremath{\mathrm{ch}}\xspace} \newcommand{\cc}{\ensuremath{\mathrm{cc}}\xspace} \newcommand{\hs}[1][]{{\frak{hs}_{#1}}} \newcommand{\Alg}{\textsl{Alg}\xspace} \newcommand{\NNO}{\mathbb{N}} \newcommand{\Z}{\mathbb{Z}} \newcommand{\R}{\mathbb{R}} \newcommand{\C}{\mathbb{C}} \newcommand{\T}{\mathbb{T}} \renewcommand{\H}{\ensuremath{\mathcal{H}}} \newcommand{\calF}{\mathcal{F}} \newcommand{\supp}{\textsl{supp}} \newcommand{\open}{\ensuremath{\mathcal{O}}} \newcommand{\closed}{\ensuremath{\mathcal{C}{\kern-.45ex}\ell}} \newcommand{\ket}[1]{\ensuremath{|{\kern.1em}#1{\kern.1em}\rangle}} \newcommand{\bigket}[1]{\ensuremath{\big|#1\big\rangle}} \newcommand{\sotimes}{\mathrel{\raisebox{.05pc}{$\scriptstyle \otimes$}}} \newcommand{\Kl}{\mathcal{K}{\kern-.5ex}\ell} \newcommand{\set}[2]{\{#1\;|\;#2\}} \newcommand{\setin}[3]{\{#1\in#2\;|\;#3\}} \newcommand{\conjun}{\mathrel{\wedge}} \newcommand{\disjun}{\mathrel{\vee}} \newcommand{\andthen}{\ensuremath{\mathrel{\&}}} \newcommand{\all}[2]{\forall{#1}.\,#2} \newcommand{\allin}[3]{\forall{#1\in#2}.\,#3} \newcommand{\ex}[2]{\exists{#1}.\,#2} \newcommand{\exin}[3]{\exists{#1\in#2}.\,#3} \newcommand{\lam}[2]{\lambda#1.\,#2} \newcommand{\lamin}[3]{\lambda#1\in#2.\,#3} \newcommand{\tuple}[1]{\langle#1\rangle} \newcommand{\charac}[1]{\mathbf{1}_{#1}} \newcommand{\intd}{{\kern.2em}\mathrm{d}{\kern.03em}} \newcommand{\indic}[1]{\mathbf{1}_{#1}} \newcommand{\downset}{\mathop{\downarrow}} \newcommand{\upset}{\mathop{\uparrow}} \newcommand{\downbit}{\ket{\downset}} \newcommand{\upbit}{\ket{\upset}} \newcommand{\Prob}[1]{\mathit{Prob}_{#1}} \newcommand{\dst}{\ensuremath{\mathsf{dst}}} \newcommand{\st}{\ensuremath{\mathsf{st}}} \newcommand{\leftScottint}{[{\kern-.3ex}[} \newcommand{\rightScottint}{]{\kern-.3ex}]} \newcommand{\Scottint}[1]{\leftScottint\,#1\,\rightScottint} \newcommand{\congrightarrow}{\mathrel{\stackrel{ \raisebox{.5ex}{$\scriptstyle\cong\,$}}{ \raisebox{0ex}[0ex][0ex]{$\rightarrow$}}}} \newcommand{\conglongrightarrow}{\mathrel{\stackrel{ \raisebox{.5ex}{$\scriptstyle\cong\,$}}{ \raisebox{0ex}[0ex][0ex]{$\longrightarrow$}}}} \newcommand{\partot}{\ensuremath{\mathcal{T}\hspace{-3pt}{\scriptstyle{o}}}} \newcommand{\totpar}{\ensuremath{\mathcal{P}\hspace{-3pt}{\scriptstyle{a}}}} \newcommand{\partotscript}{\ensuremath{\mathcal{T}\hspace{-3pt}{\scriptscriptstyle{o}}}} \newcommand{\totparscript}{\ensuremath{\mathcal{P}\hspace{-2pt}{\scriptscriptstyle{a}}}} \newcommand{\pth}{\ensuremath{\hat{\mathcal{T}}\hspace{-3pt}{\scriptstyle{o}}}} \newcommand{\tph}{\ensuremath{\hat{\mathcal{P}}\hspace{-3pt}{\scriptstyle{a}}}} \newcommand{\ojoin}{\ovee} \newcommand{\eps}{\varepsilon} \newcommand{\marg}[2]{\ensuremath{{#1}\mathrel{\%}{#2}}} \newcommand{\margfun}{\mathsf{M}} \newcommand{\betachan}{\ensuremath{\mathrm{Beta}}} \newcommand{\flipchan}{\ensuremath{\mathrm{Flip}}} \newcommand{\normchan}{\ensuremath{\mathrm{Norm}}} \newcommand{\binomchan}{\ensuremath{\mathrm{Binom}}} \newcommand{\dirchan}{\ensuremath{\mathrm{Dir}}} \newcommand{\multchan}{\ensuremath{\mathrm{Mult}}} \newcommand{\copychan}{\ensuremath{\mathrm{copy}}} \newcommand{\LHS}{\ensuremath{\mathrm{LHS}}} \newcommand{\RHS}{\ensuremath{\mathrm{RHS}}} \newsavebox\sbground \savebox\sbground{\begin{tikzpicture}[baseline=0pt,circuit ee IEC] \draw (0,-.1ex) to (0,.85ex) node[ground,point up,scale=.6,xshift=.55ex] {}; \end{tikzpicture}} \newcommand{\ground}{\mathord{\usebox\sbground}} \newsavebox\sbcopier \savebox\sbcopier{\begin{tikzpicture}[baseline=0pt] \node[copier,scale=.7] (a) at (0,3.6pt) {}; \draw (a) -- +(-90:.16); \draw (a) -- +(45:.19); \draw (a) -- +(135:.19); \end{tikzpicture}} \newcommand{\copier}{\mathord{\usebox\sbcopier}} \DeclareFixedFont{\ttb}{T1}{txtt}{bx}{n}{12} \DeclareFixedFont{\ttm}{T1}{txtt}{m}{n}{12} \usepackage{color} \definecolor{deepblue}{rgb}{0,0,0.5} \definecolor{deepred}{rgb}{0.6,0,0} \definecolor{deepgreen}{rgb}{0,0.5,0} \newcommand\pythonstyle{\lstset{ backgroundcolor = \color{lightgray}, language=Python, basicstyle=\ttm, otherkeywords={self,>>>}, keywordstyle=\ttb\small\color{deepblue}, emph={@,+\%literate={.+}{{{\color{red}.+}}}2 {.**}{{{\color{red}.\**{}}}}2 {*}{{{\color{red}*}}}1 }, emphstyle=\ttb\small\color{deepred}, stringstyle=\small\color{deepgreen}, frame=tb, showstringspaces=false }} \lstnewenvironment{python}[1][] { \pythonstyle \lstset{basicstyle=\normalfont\ttfamily\small#1} } {} \newcommand\pythonexternal[2][]{{ \pythonstyle \lstinputlisting[#1]{#2}}} \newcommand\pythoninline[1]{\pythonstyle\lstinline[basicstyle=\normalfont\ttfamily\small]{#1}} \newtheorem{theorem}{Theorem}[section] \newtheorem{proposition}[theorem]{Proposition} \newtheorem{lemma}[theorem]{Lemma} \newtheorem{corollary}[theorem]{Corollary} \newtheorem{definition}[theorem]{Definition} \newtheorem{remark}[theorem]{Remark} \newtheorem{example}[theorem]{Example} \newtheorem{notation}[theorem]{Notation} \renewcommand{\arraystretch}{1.3} \setlength{\arraycolsep}{3pt} \renewcommand{\thefootnote}{\arabic{footnote}} \bibliographystyle{apalike} \title[A Channel-Based Perspective on Conjugate Priors]{A Channel-Based Perspective on Conjugate Priors\thanks{The research leading to these results has received funding from the European Research Council under the European Union's Seventh Framework Programme (FP7/2007-2013) / ERC grant agreement nr.~320571}} \author[B. Jacobs]{B\ls A\ls R\ls T\ns J\ls A\ls C\ls O\ls B\ls S\\ Institute for Computing and Information Sciences Radboud University\addressbreak P.O.Box 9010, 6500 GL Nijmegen, the Netherlands} \begin{document} \maketitle \begin{abstract} A desired closure property in Bayesian probability is that an updated posterior distribution be in the same class of distributions --- say Gaussians --- as the prior distribution. When the updating takes place via a statistical model, one calls the class of prior distributions the `conjugate priors' of the model. This paper gives (1)~an abstract formulation of this notion of conjugate prior, using channels, in a graphical language, (2)~a simple abstract proof that such conjugate priors yield Bayesian inversions, and (3)~a logical description of conjugate priors that highlights the required closure of the priors under updating. The theory is illustrated with several standard examples, also covering multiple updating. \end{abstract} \section{Introduction}\label{sec:intro} The main result of this paper, Theorem~\ref{thm:conjugateinversion}, is mathematically trivial. But it is not entirely trivial to see that this result is trivial. The effort and contribution of this paper lies in setting up a framework --- using the abstract language of channels, Kleisli maps, and string diagrams for probability theory --- to define the notion of conjugate prior in such a way that there is a trivial proof of the main statement, saying that conjugate priors yield Bayesian inversions. This is indeed what conjugate priors are meant to be. Conjugate priors form a fundamental topic in Bayesian theory. They are commonly described via a closure property of a class of prior distributions, namely as being closed under certain Bayesian updates. Conjugate priors are especially useful because they do not only involve a closure \emph{property}, but also a particular \emph{structure}, namely an explicit function that performs an analytical computation of posterior distributions via updates of the parameters. This greatly simplify Bayesian analysis. For instance, the $\betachan$ distribution is conjugate prior to the Bernoulli (or `flip') distribution, and also to the binomial distribution: updating a $\betachan(\alpha,\beta)$ prior via a Bernoulli/binomial statistical model yields a new $\betachan(\alpha',\beta')$ prior, with adapted parameters $\alpha',\beta'$ that can be computed explicitly from $\alpha,\beta$ and the observation at hand. Despite this importance, the descriptions in the literature of what it means to be a conjugate prior are remarkably informal. One does find several lists of classes of distributions, for instance at Wikipedia\footnote{See \url{https://en.wikipedia.org/wiki/Conjugate_prior} or online lists, such as \url{https://www.johndcook.com/CompendiumOfConjugatePriors.pdf}, consulted at Sept.\ 10, 2018}, together with formulas about how to re-compute parameters. The topic has a long and rich history in statistics (see \textit{e.g.}~\cite{Bishop06}), with much emphasis on exponential families~\cite{DiaconisY79}, but a precise, general definition is hard to find. We briefly review some common approaches, without any pretension to be complete: the definition in~\cite{Alpaydin10} is rather short, based on an example, and just says: ``We see that the posterior has the same form as the prior and we call such a prior a \emph{conjugate prior}.'' Also~\cite{RussellN03} mentions the term `conjugate prior' only in relation to an example. There is a separate section in~\cite{Bishop06} about conjugate priors, but no precise definition. Instead, there is the informal description ``\ldots the posterior distribution has the same functional form as the prior.'' The most precise definition (known to the author) is in~\cite[\S5.2]{BernardoS00}, where the conjugate family with respect to a statistical model, assuming a `sufficient statistic', is described. It comes close to our channel-based description, since it explicitly mentions the conjugate family as a conditional probability distribution with (re-computed) parameters. The approach is rather concrete however, and the high level of mathematical abstraction that we seek here is missing in~\cite{BernardoS00}. This paper presents a novel systematic perspective for precisely defining what conjugate priorship means, both via diagrams and via (probabilistic) logic. It uses the notion of `channel' as starting point. The basis of this approach lies in category theory, especially effectus theory~\cite{Jacobs15d,ChoJWW15b}. However, we try to make this paper accessible to non category theorists, by using the term `channel' instead of morphism in a Kleisli category of a suitable monad. Moreover, a graphical language is used for channels that hopefully makes the approach more intuitive. Thus, the emphasis of the paper is on \emph{what it means} to have conjugate priors. It does not offer new perspectives on how to find/obtain them. The paper is organised as follows. It starts in Section~\ref{sec:ideas} with a high-level description of the main ideas, without going into technical details. Preparatory definitions are provided in Sections~\ref{sec:Kleisli} and~\ref{sec:inversion}, dealing with channels in probabilistic computation, with a diagrammatic language for channels, and with Bayesian inversion. Then, Section~\ref{sec:conjugate} contains the novel channel-based definition of conjugate priorship; it also illustrates how several standard examples fit in this new setting. Section~\ref{sec:conjugatepriorinversion} establishes the (expected) close relationship between conjugate priors and Bayesian inversions. Section~\ref{sec:logic} then takes a fresh perspective, by re-describing the Bayesian-inversion based formulation in more logical terms, using validity and updating. This re-formulation captures the intended closure of a class of priors under updating in the most direct manner. It is used in Section~\ref{sec:multiple} to illustrate how multiple updates are handled, typically via a `sufficient statistic'. \section{Main ideas}\label{sec:ideas} This section gives an informal description of the main ideas underlying this paper. It starts with a standard example, and then proceeds with a step-by-step introduction to the essentials of the perspective of this paper. \begin{figure} \begin{center} \includegraphics[width=0.3\textwidth]{coin.eps} \includegraphics[width=0.3\textwidth]{coin-H.eps} \includegraphics[width=0.3\textwidth]{coin-HTTT.eps} \end{center} \caption{Uniform prior, and two posterior probability density functions on $[0,1]$, after observing head, and after observing head-tail-tail-tail. These functions correspond respectively to $\betachan(1,1)$, $\betachan(2,1)$, $\betachan(2,4)$. Example~\ref{ex:efprobcoin} below explains how these three plots are obtained, via actual Bayesian updates (inversions), and not by simply using the $\betachan$ functions.} \label{fig:coin} \end{figure} A well-known example in Bayesian reasoning is inferring the (unknown) bias of a coin from a sequence of consecutive head/tail observations. The bias is a number $r \in [0,1]$ in the unit interval, giving the `Bernoulli' or `flip' probability $r$ for head, and $1-r$ for tail. Initially we assume a uniform distribution for $r$, as described by the constant probability density function (pdf) on the left in Figure~\ref{fig:coin}. After observing one head, this pdf changes to the second picture. After observing head-tail-tail-tail we get the third pdf. These pictures are obtained by Bayesian inversion, see Section~\ref{sec:inversion}. It is a well-known fact that all the resulting distributions are instances of the $\betachan(\alpha,\beta)$ family of distributions, for different parameters $\alpha,\beta$. After each observation, one can re-compute the entire updated distribution, via Bayesian inversion, as in Example~\ref{ex:efprobcoin}. But in fact there is a much more efficient way to obtain the revised distribution, namely by computing the new parameter values: increment $\alpha$ by one, for head, and increment $\beta$ by one for tail, see Examples~\ref{ex:betaflip} and~\ref{ex:betaflipupdate} for details. The family of distributions $\betachan(\alpha,\beta)$, indexed by parameters $\alpha,\beta$, is thus suitably closed under updates with Bernoulli. It is the essence of the statement that $\betachan$ is conjugate prior to Bernoulli. This will be made precise later on. Let $X = (X, \Sigma)$ be a measurable space, where $\Sigma \subseteq \Pow(X)$ is a $\sigma$-algebra of measurable subsets. We shall write $\Giry(X)$ for the set of probability distributions on $X$. Elements $\omega\in\Giry(X)$ are thus countably additive functions $\omega\colon\Sigma \rightarrow [0,1]$ with $\omega(X) = 1$. \begin{description} \item[Idea 1:] A \emph{family} of distributions on $X$, indexed by a measurable space $P$ of parameters, is a (measurable) function $P \rightarrow \Giry(X)$. Categorically, such a function is a \emph{Kleisli} map for $\Giry$, considered as monad on the category of measurable spaces (see Section~\ref{sec:Kleisli}). These Kleisli maps are also called \emph{channels}, and will be written simply as arrows $P \rightarrow X$, or diagrammatically as boxes \ $\scriptstyle\vcenter{\hbox{\begin{tikzpicture}[font=\tiny] \node[arrow box, scale=0.5] (c) at (0,0) {$\;$}; \draw (c) to (0,0.35); \draw (c) to (0,-0.35); \node at (0.2,0.3) {$X$}; \node at (0.2,-0.3) {$P$}; \end{tikzpicture}}}$ where we imagine that information is flowing upwards. \end{description} \noindent The study of families of distributions goes back a long way, \textit{e.g.} as `experiments'~\cite{Blackwell51}. Along these lines we shall describe the family of $\betachan$ distributions as a channel with $P = \R_{>0}\times\R_{>0}$ and $X = [0,1]$, namely as function: \begin{equation} \label{diag:beta} \vcenter{\xymatrix@C+1pc{ \R_{>0}\times\R_{>0}\ar[r]^-{\betachan} & \Giry([0,1]) }} \end{equation} \noindent For $(\alpha,\beta) \in \R_{>0}\times\R_{>0}$ there is the probability distribution $\betachan(\alpha,\beta) \in \Giry([0,1])$ determined by its value on a measurable subset $M\subseteq [0,1]$, which is obtained via integration: \begin{equation} \label{eqn:beta} \begin{array}{rcl} \betachan(\alpha,\beta)(M) & = & {\displaystyle\int_{M}} \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)}\intd x, \end{array} \end{equation} \noindent where $B(\alpha,\beta) = \int_{[0,1]} x^{\alpha-1}(1-x)^{\beta-1} \intd x$ is a normalisation constant. A conjugate prior relationship involves a family of distributions $P \rightarrow \Giry(X)$ which is closed wrt.\ updates based on observations (or: data) from a separate domain $O$. Each `parameter' element $x\in X$ gives rise to a separate distribution on $O$. This is what is usually called a \emph{statistical} or \emph{parametric} model. We shall also describe it as a channel. \begin{description} \item[Idea 2:] The observations for a family $P \rightarrow \Giry(X)$ arise via another ``Kleisli'' map $X \rightarrow \Giry(O)$ representing the statistical model. Conjugate priorship will be defined for two such composable channels $P \rightarrow X \rightarrow O$, where $O$ is the space of observations. \end{description} In the above coin example, the space $O$ of observations is the two-element set $2 = \{0,1\}$ where $0$ is for tail and $1$ for head. The Bernoulli channel is written as $\flipchan \colon [0,1] \rightarrow \Giry(2)$. A probability $r\in [0,1]$ determines a Bernoulli/flip/coin probability distribution $\flipchan(r) \in \Giry(2)$ on $2$, formally sending the subset $\{1\}$ to $r$ and $\{0\}$ to $1-r$. \begin{description} \item[Idea 3:] A channel $c\colon P\rightarrow X$ is a conjugate prior to a channel $d\colon X \rightarrow O$ if there is a \emph{parameter translation function} $h\colon P\times O \rightarrow P$ satisfying a suitable equation. \end{description} \noindent The idea is that $c(p)$ is a prior, for $p\in P$, which gets updated via the statistical model (channel) $d$, in the light of observation $y\in O$. The revised, updated distribution is $c(h(p,y))$. The model $d$ is usually written as a conditional probability $d(y\mid \theta)$. In the coin example we have $h\colon \R_{>0}\times\R_{>0} \times 2 \rightarrow \R_{>0}\times\R_{>0}$ given by $h(\alpha,\beta,1) = (\alpha+1,\beta)$ and $h(\alpha,\beta,0) = (\alpha,\beta+1)$, see Example~\ref{ex:betaflip} below for more information. What has been left unexplained is the `suitable' equation that the parameter translation function $h\colon P\times O \rightarrow P$ should satisfy. It is not entirely trivial, because it is an equation between channels in what is called the Kleisli category $\Kl(\Giry)$ of the Giry monad $\Giry$. At this stage we need to move to a more categorical description. The equation, which will appear in Definition~\ref{def:conjugateprior}, bears similarities with the notion of Bayesian inversion, which will be introduced in Section~\ref{sec:inversion}. \section{Channels and conditional probabilities}\label{sec:Kleisli} This section will describe conditional probabilities as arrows and will show how to compose them. Thereby we are entering the world of category theory. We aim to suppress the underlying categorical machinery and make this work accessible to readers without such background. For those with categorical background knowledge: we will be working in the Kleisli categories of the distribution monad $\Dst$ for discrete probability, and of the Giry monad $\Giry$ for continuous probability, see \textit{e.g.}~\cite{Giry82,Panangaden09,Jacobs17a}. Discrete distributions may be seen as a special case of continuous distributions, via a suitable inclusion map $\Dst \rightarrow \Giry$. Hence one could give one account, using $\Giry$ only. However, in computer science, unlike for instance in statistics, discrete distributions are so often used that they merit separate treatment. We thus start with discrete probability. We write a (finite, discrete) distribution on a set $X$ as a formal convex sum $r_{1}\ket{x_1} + \cdots + r_{n}\ket{x_n}$ of elements $x_{i}\in X$ and probabilities $r_{i}\in [0,1]$ with $\sum_{i}r_{i}=1$. The `ket' notation $\ket{-}$ is syntactic sugar, used to distinguish elements of $x$ from their occurrence $\ket{x}$ in such formal convex sums\footnote{Sometimes these distributions $\sum_{i}r_{i}\ket{x_i}$ are called `multinomial' or `categorical'; the latter terminology is confusing in the present context.}. A distribution as above can be identified with a `probability mass' function $\omega \colon X \rightarrow [0,1]$ which is $r_{i}$ on $x_{i}$ and $0$ elsewhere. We often implicitly identify distributions with such functions. We shall write $\Dst(X)$ for the set of distributions on $X$. We shall focus on functions of the form $c\colon X \rightarrow \Dst(Y)$. They give, for each element $x\in X$ a distribution $c(x)$ on $Y$. Hence such functions form an $X$-indexed collection $\big(c(x)\big)_{x\in X}$ of distributions $c(x)$ on $Y$. They can be understood as \emph{conditional} probabilities $P(y\mid x) = r$, if $c(x)$ is of the form $\cdots r\ket{y}\cdots$, with weight $r = c(x)(y)\in[0,1]$ for $y\in Y$. Thus, by construction, $\sum_{y} P(y\mid x) = 1$, for each $x\in X$. Moreover, if the sets $X$ and $Y$ are finite, we can describe $c\colon X \rightarrow \Dst(Y)$ as a stochastic matrix, with entries $P(y\mid x)$, adding up to one --- per row or column, depending on the chosen orientation of the matrix. We shall often write functions $X \rightarrow \Dst(Y)$ simply as arrows $X \rightarrow Y$, call them `channels', and write them as `boxes' in diagrams. This arrow notation is justified, because there is a natural way to compose channels, as we shall see shortly. But first we describe \emph{state transformation}, also called \emph{prediction}. Given a channel $c\colon X \rightarrow \Dst(Y)$ and a state $\omega\in\Dst(X)$, we can form a new state, written as $c \gg \omega$, on $Y$. It is defined as: \begin{equation} \label{eqn:discstatransf} \begin{array}{rcl} c \gg \omega & \coloneqq & {\displaystyle\sum_y} \big(\sum_{x} \omega(x)\cdot c(x)(y)\big)\bigket{y}. \end{array} \end{equation} \noindent The outer sum $\sum_{y}$ is a formal convex sum, whereas the inner sum $\sum_{x}$ is an actual sum in the unit interval $[0,1]$. Using state transformation $\gg$ it is easy to define composition of channels: given functions $c\colon X \rightarrow \Dst(Y)$ and $d\colon Y \rightarrow \Dst(Z)$, we use the ordinary composition symbol $\after$ to form a composite channel $d \after c \colon X \rightarrow \Dst(Z)$, where: \begin{equation} \label{eqn:discretecomposition} \begin{array}{rcccl} (d \after c)(x) & \coloneqq & d \gg c(x) & = & {\displaystyle\sum_{z\in Z}} \big(\sum_{y} c(x)(y)\cdot d(y)(z)\big) \bigket{z}. \end{array} \end{equation} \noindent Essentially, this is matrix composition for stochastic matrices. Channel composition $\after$ is associative, and also has a neutral element, namely the identity channel $\eta \colon X \rightarrow X$ given by the `Dirac' function $\eta(x) = 1\ket{x}$. It is not hard to see that $(d \after c) \gg \omega = d \gg (c \gg \omega)$. \medskip We turn to channels in continuous probability. As already mentioned in Section~\ref{sec:ideas}, we write $\Giry(X)$ for the set of probability distributions $\omega\colon \Sigma_{X} \rightarrow [0,1]$, where $X = (X,\Sigma_{X})$ is a measurable space. These probability distributions are (also) called states. The set $\Giry(X)$ carries a $\sigma$-algebra itself, but that does not play an important role here. Each element $x\in X$ yields a probability measure $\eta(x)\in\Giry(X)$, with $\eta(x)(M) = \indic{M}(x)$, which is $1$ if $x\in M$ and $0$ otherwise. This map $\indic{M} \colon X \rightarrow [0,1]$ is called the indicator function for the subset $M\in\Sigma_{X}$. For a state/measure $\omega\in\Giry(X)$ and a measurable function $f\colon X \rightarrow \R_{\geq 0}$ we write $\int f\intd \omega$ for the Lebesgue integral, if it exists. We follow the notation of~\cite{Jacobs13a} and refer there for details, or alternatively, to~\cite{Panangaden09}. We recall that an integral $\int_{M} f\intd \omega$ over a measurable subset $M\subseteq X$ of the domain of $f$ is defined as $\int \indic{M}\cdot f \intd\omega$, and that $\int \indic{M} \intd\omega = \omega(M)$. Moreover, $\int f \intd\eta(x) = f(x)$. For a measurable function $g\colon X\rightarrow Y$ between measurable spaces $X,Y$ there is the `push forward' function $\Giry(g) \colon \Giry(X) \rightarrow \Giry(Y)$, given by $\Giry(g)(\omega)(N) = \omega\big(g^{-1}(N)\big)$. It satisfies: \begin{equation} \label{eqn:invmeasure} \begin{array}{rcl} \displaystyle\int f \intd \Giry(g)(\omega) & = & \int f \after g \intd \omega. \end{array} \end{equation} \noindent Often, the measurable space $X$ is a subset $X\subseteq\R$ of the real numbers and a probability distribution $\omega$ on $X$ is given by a probability density function (pdf), that is, by a measurable function $f\colon X \rightarrow \R_{\geq 0}$ with $\int_{X} f(x) \intd x = 1$. Such a pdf $f$ gives rise to a state $\omega \in\Giry(X)$, namely: \begin{equation} \label{eqn:statefrompdf} \begin{array}{rcl} \omega(M) & = & \displaystyle \int_{M} f(x) \intd x. \end{array} \end{equation} \noindent We then write $\omega = \int f$. In this continuous context a channel is a measurable function $c\colon X \rightarrow \Giry(Y)$, for measurable spaces $X,Y$. Like in the discrete case, it gives an $X$-indexed collection $(\big(c(x)\big)_{x\in X}$ of probability distributions on $Y$. The channel $c$ can transform a state $\omega\in\Giry(X)$ on $X$ into a state $c \gg \omega \in\Giry(Y)$ on $Y$, given on a measurable subset $N\subseteq Y$ as: \begin{equation} \label{eqn:statetransformation} \begin{array}{rcl} \big(c \gg \omega\big)(N) & = & \displaystyle\int c(-)(N) \intd \omega. \end{array} \end{equation} \noindent For another channel $d\colon Y \rightarrow \Giry(Z)$ there is a composite channel $d \after c \colon X \rightarrow \Giry(Z)$, via integration: \begin{equation} \label{eqn:continuouscomposition} \begin{array}{rcccl} \big(d \after c\big)(x)(K) & \coloneqq & \big(d \gg c(x)\big)(K) & = & \displaystyle\int d(-)(K) \intd c(x) \end{array} \end{equation} In many situations a channel $c\colon X \rightarrow \Giry(Y)$ is given by an indexed probability density function (pdf) $u\colon X \times Y \rightarrow \R_{\geq 0}$, with $\int u(x,y)\intd y = 1$ for each $x\in X$. The associated channel $c$ is: \begin{equation} \label{eqn:channelfrompdf} \begin{array}{rcl} c(x)(N) & = & \displaystyle \int_{N} u(x,y) \intd y. \end{array} \end{equation} \noindent In that case we simply write $c = \int u$ and call $c$ a pdf-channel. We have already seen such a description of the $\betachan$ distribution as a pdf-channel in~\eqref{eqn:beta}. (In these pdf-channels $X \rightarrow Y$ we use a collection of pdf's $u(x,-)$ which are all dominated by the Lebesgue measure. This domination happens via the relationship $\ll$ of absolute continuity, using the Radon-Nikodym Theorem, see \textit{e.g.}~\cite{Panangaden09}.) Various additional computation rules for integrals are given in the Appendix. \section{Bayesian inversion in string diagrams}\label{sec:inversion} In this paper we make superficial use of string diagrams to graphically represent sequential and parallel composition of channels, mainly in order to provide an intuitive visual overview. We refer to~\cite{Selinger11} for mathematical details, and mention here only the essentials. A channel $X \rightarrow Y$, for instance of the sort discussed in the previous section, can be written as a box \ $\scriptstyle\vcenter{\hbox{\begin{tikzpicture}[font=\tiny] \node[arrow box, scale=0.5] (c) at (0,0) {$\;$}; \draw (c) to (0,0.35); \draw (c) to (0,-0.35); \node at (0.2,0.3) {$Y$}; \node at (0.2,-0.3) {$X$}; \end{tikzpicture}}}$ with information flowing upwards, from the wire labeled with $X$ to the wire labeled with $Y$. Composition of channels, as in~\eqref{eqn:discretecomposition} or~\eqref{eqn:continuouscomposition}, simply involves connecting wires (of the same type). The identity channel is just a wire. We use a triangle notation \ \raisebox{.3em}{$\scriptstyle\vcenter{\hbox{\begin{tikzpicture}[font=\tiny] \node[state, scale=0.5] (omega) at (0,0) {$\;$}; \draw (omega) to (0,0.2); \node at (0.15,0.15) {$X$}; \end{tikzpicture}}}$} for a state on $X$. It is special case of a channel, namely of the form $1 \rightarrow X$ with trivial singleton domain $1$. In the present (probabilistic) setting we allow copying of wires, written diagrammatically as $\copier$. We briefly describe such copy channels for discrete and continuous probability: \[ \xymatrix@R-1.8pc{ X\ar[r]^-{\copier} & \Dst(X\times X) & & X\ar[r]^-{\copier} & \Giry(X\times X) \\ x\ar@{|->}[r] & 1\ket{x,x} & & x\ar@{|->}[r] & \big(M\times N \mapsto \indic{M\cap N}(x)\big) } \] After such a copy we can use parallel channels. We briefly describe how this works, first in the discrete case. For channels $c\colon X \rightarrow \Dst(Y)$ and $d\colon A \rightarrow \Dst(B)$ we have a channel $c\otimes d \colon X\times A \rightarrow \Dst(Y\times B)$ given by: \[ \begin{array}{rcl} (c\otimes d)(x,a) & = & \displaystyle \sum_{y,b} c(x)(y)\cdot d(a)(b)\bigket{y,b}. \end{array} \] Similarly, in the continuous case, for channels $c\colon X \rightarrow \Giry(Y)$ and $d\colon A \rightarrow \Giry(B)$ we get $c\otimes d \colon X\times A \rightarrow \Giry(Y\times B)$ given by: \[ \begin{array}{rcl} (c\otimes d)(x,a)(M\times N) & = & c(x)(M)\cdot d(a)(N). \end{array} \] \noindent Recall that the product measure on $Y\times B$ is generated by measurable rectangles of the form $M\times N$, for $M\in\Sigma_{Y}$ and $N\in\Sigma_{B}$. We shall use a tuple $\tuple{c,d}$ as convenient abbreviation for $(c\otimes d) \after \copier$. Diagrammatically, parallel channels are written as adjacent boxes. We can now formulate what Bayesian inversion is. The definition is couched in purely diagrammatic language, but is applied only to probabilistic interpretations in this paper. \begin{definition} \label{def:inversion} The \emph{Bayesian inversion} of a channel $c\colon X \rightarrow Y$ with respect to a state $\omega$ of type $X$, if it exists, is a channel in the opposite direction, written as $c^{\dag}_{\omega} \colon Y \rightarrow X$, such that the following equation holds. \begin{equation} \label{diag:inversion} \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (omega) at (0,0) {$\omega$}; \node[copier] (copier) at (0,0.3) {}; \node[arrow box] (c) at (0.5,0.95) {$c$}; \coordinate (X) at (-0.5,1.5); \coordinate (Y) at (0.5,1.5); \draw (omega) to (copier); \draw (copier) to[out=150,in=-90] (X); \draw (copier) to[out=15,in=-90] (c); \draw (c) to (Y); \end{tikzpicture}}} \qquad=\qquad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (omega) at (0,-0.55) {$\omega$}; \node[copier] (copier) at (0,0.3) {}; \node[arrow box] (d) at (-0.5,0.95) {$c^{\dag}_{\omega}$}; \coordinate (X) at (-0.5,1.5); \coordinate (Y) at (0.5,1.5); \node[arrow box] (c) at (0,-0.15) {$c$}; \draw (omega) to (c); \draw (c) to (copier); \draw (copier) to[out=165,in=-90] (d); \draw (copier) to[out=30,in=-90] (Y); \draw (d) to (X); \end{tikzpicture}}} \end{equation} \end{definition} The dagger notation $c^{\dag}_{\omega}$ is copied from~\cite{ClercDDG17}. There the state $\omega$ is left implicit, via a restriction to a certain comma category of kernels. In that setting the operation $(-)^{\dag}$ is functorial, and forms a dagger category (see \textit{e.g.}~\cite{AbramskyC09,Selinger07} for definitions). In particular, it preserves composition and identities of channels. Equation~\eqref{diag:inversion} can also be written as: $\tuple{\idmap, c} \gg \omega = \tuple{c^{\dag}_{\omega}, \idmap} \gg (c \gg \omega)$. Alternatively, in the discrete case, with variables explicit, it says: $c(x)(y)\cdot \omega(x) = c^{\dag}_{\omega}(y)(x) \cdot (c \gg \omega)(y)$. This comes close to the `adjointness' formulations that are typical for daggers. Bayesian inversion gives a channel-based description of Bayesian (belief) updates. We briefly illustrate this for the coin example from Section~\ref{sec:ideas}, using the $\EfProb$ language~\cite{ChoJ17b}. \begin{example} \label{ex:efprobcoin} In Section~\ref{sec:ideas} we have seen the channel $\flipchan \colon [0,1] \rightarrow 2$ that sends a probability $r\in [0,1]$ to the coin state $\flipchan(r) = r\ket{1} + (1-r)\ket{0}$ with bias $r$. The Bayesian inversion $2 \rightarrow [0,1]$ of this channel performs a belief update, after a head/tail observation. Without going into details we briefly illustrate how this works in the $\EfProb$ language via the following code fragment. The first line describes a channel \pythoninline{Flip} of type $[0,1] \rightarrow 2$, where $[0,1]$ is represented as \pythoninline{R(0,1)} and $2 = \{0,1\}$ as \pythoninline{bool\_dom}. The expression \pythoninline{flip(r)} captures a coin with bias \pythoninline{r}. \begin{python} >>> Flip = chan_fromklmap(lambda r: flip(r), R(0,1), bool_dom) >>> prior = uniform_state(R(0,1)) >>> w1 = Flip.inversion(prior)(True) >>> w2 = Flip.inversion(w1)(False) >>> w3 = Flip.inversion(w2)(False) >>> w4 = Flip.inversion(w3)(False) \end{python} \noindent The (continuous) states \pythoninline{w1} -- \pythoninline{w4} are obtained as successive updates of the uniform state \pythoninline{prior}, after successive observations \pythoninline{True}-\pythoninline{False}-\pythoninline{False}-\pythoninline{False}, for head-tail-tail-tail. The three probability density functions in Figure~\ref{fig:coin} are obtained by plotting the prior state, and also the two states \pythoninline{w1} and \pythoninline{w4}. \end{example} It is relatively easy to define Bayesian inversion in discrete probability theory: for a channel $c\colon X \rightarrow \Dst(Y)$ and a state/distribution $\omega\in\Dst(X)$ one can define a channel $c^{\dag}_{\omega} \colon Y \rightarrow \Dst(X)$ as: \begin{equation} \label{eqn:discreteinversion} \begin{array}{rcccl} c^{\dag}_{\omega}(y)(x) & = & \displaystyle\frac{\omega(x)\cdot c(x)(y)}{(c \gg \omega)(y)} & = & \displaystyle\frac{\omega(x)\cdot c(x)(y)}{\sum_{z}\omega(z)\cdot c(z)(y)}, \end{array} \end{equation} \noindent assuming that the denominator is non-zero. This corresponds to the familiar formula $P(B\mid A) = \nicefrac{P(A,B)}{P(A)}$ for conditional probability. The state $c^{\dag}_{\omega}(y)$ can alternatively be defined via updating the state $\omega$ with the point predicate $\{y\}$, transformed via $c$ into a predicate $c \ll \indic{\{y\}}$ on $X$, see Section~\ref{sec:logic} (and~\cite{JacobsZ16}) for details. \begin{Auxproof} Just to be sure: \[ \begin{array}{rcl} \big(\tuple{c^{\dag}_{\omega}, \idmap} \gg (c \gg \omega)\big)(x,y) & = & c^{\dag}_{\omega}(y)(x)\cdot (c \gg \omega)(y) \\ & = & \displaystyle\frac{\omega(x)\cdot c(x)(y)}{c_{*}(\omega)(y)} \cdot (c \gg \omega)(y) \\ & = & \omega(x)\cdot c(x)(y) \\ & = & (\tuple{\idmap,c} \gg \omega)(x,y). \end{array} \] \noindent And: \[ \begin{array}{rcccl} \omega|_{c \ll \indic{\{y\}}}(x) & = & \displaystyle\frac{\omega(x) \cdot (c \ll \indic{\{y\}})(x)}{\omega \models c \ll \indix{\{y\}}} & = & \displaystyle\frac{\omega(x)\cdot c(x)(y)}{\sum_{x}\omega(x)\cdot c(x)(y)}, \end{array} \] \end{Auxproof} The situation is much more difficult in continuous probability theory, since Bayesian inversions may not exist~\cite{AckermanFR11,Stoyanov14} or may be determined only up to measure zero. But when restricted to \textit{e.g.}~standard Borel spaces, as in~\cite{ClercDDG17}, existence is ensured, see also~\cite{Faden85,CulbertsonS14}. Another common solution is to assume that we have a pdf-channel: there is a map $u\colon X\times Y \rightarrow \R_{\geq 0}$ that defines a channel $c \colon X \rightarrow \Giry(Y)$, like in~\eqref{eqn:channelfrompdf}, as $c(x)(N) = \int_{N} u(x,y)\intd y$. Then, for a distribution $\omega\in\Giry(X)$ we can take as Bayesian inversion: \begin{equation} \label{eqn:continuousinversion} \begin{array}{rcl} c^{\dag}_{\omega}(y)(M) & = & \displaystyle\frac{\int_{M} u(-,y)\intd \omega}{\int_{X} u(-,y)\intd \omega} \\[+1em] & = & \displaystyle\frac{\int_{M} f(x)\cdot u(x,y) \intd x} {\int_{X} f(x) \cdot u(x,y) \intd x} \qquad \mbox{when } \omega = \int f(x)\intd x. \end{array} \end{equation} \noindent We prove that this definition satisfies the inversion Equation~\eqref{diag:inversion}, using the calculation rules from the Appendix. \[ \begin{array}{rcl} \big(\tuple{c^{\dag}_{\omega}, \idmap} \gg (c \gg \omega)\big)(M\times N) & \smash{\stackrel{\eqref{eqn:statetransformation}}{=}} & \displaystyle \int \tuple{c^{\dag}_{\omega}, \idmap}(-)(M\times N) \intd (c \gg \omega) \\ & \smash{\stackrel{(\ref{eqn:pdfintegration}, \ref{eqn:pdfstatetransformation})}{=}} & \displaystyle\int \big(\int f(x) \cdot u(x,y) \intd x\big) \cdot \tuple{c^{\dag}_{\omega}, \idmap}(y)(M\times N) \intd y \\ & \smash{\stackrel{\eqref{eqn:graphequation}}{=}} & \displaystyle\int \big(\int f(x) \cdot u(x,y) \intd x\big) \cdot c^{\dag}_{\omega}(y)(M) \cdot \indic{N}(y) \intd y \\ & \smash{\stackrel{\eqref{eqn:continuousinversion}}{=}} & \displaystyle\int_{N} \big(\int f(x) \cdot u(x,y) \intd x\big) \cdot \frac{\int_{M} f(x)\cdot u(x,y) \intd x} {\int f(x)\cdot u(x,y) \intd x} \intd y \\[+0.8em] & = & \displaystyle\int_{N} \int_{M} f(x) \cdot u(x,y) \intd x \intd y \\ & \smash{\stackrel{\eqref{eqn:pdfgraphstatetransformation}}{=}} & (\tuple{\idmap, c} \gg \omega)(M\times N). \end{array} \] \section{Conjugate priors}\label{sec:conjugate} We now come to the core of this paper. As described in the introduction, the informal definition says that a class of distributions is conjugate prior to a statistical model if the associated posteriors are \emph{in the same class} of distributions. The posteriors can be computed via Bayesian inversion~\eqref{eqn:continuousinversion} of the statistical model. This definition of `conjugate prior' is a bit vague, since it loosely talks about `classes of distributions', without further specification. As described in `Idea 1' in Section~\ref{sec:ideas}, we interpret `class of states on $X$' as channel $P\rightarrow X$, where $P$ is the type of parameters of the class. We have already seen this channel-based description for the class $\betachan$ distributions, in~\eqref{diag:beta}, as channel $\betachan \colon \R_{>0} \times \R_{>0} \rightarrow [0,1]$. This works more generally, for instance for Gaussian (normal) distributions $\normchan(\mu, \sigma)$, where $\mu$ is the mean parameter and $\sigma$ is the standard deviation parameter, giving a channel of the form: \begin{equation} \label{diag:normal} \vcenter{\xymatrix@C+1pc{ \R\times\R_{>0}\ar[r]^-{\normchan} & \Giry(\R) }} \end{equation} \noindent It is determined by its value on a measurable subset $M\subseteq \R$ as the standard integral: \begin{equation} \label{eqn:normal} \begin{array}{rcl} \normchan(\mu,\sigma)(M) & = & \displaystyle\int_{M} \frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{(x-\mu)^{2}}{2\sigma^{2}}}\intd x \end{array} \end{equation} Given a channel $c\colon P \rightarrow X$, we shall look at states $c(p)$, for parameters $p\in P$, as priors. The statistical model, for which these $c(p)$'s will be described as conjugate priors, goes from $X$ to some other object $O$ of `observations'. Thus our starting point is a pair of (composable) channels the form: \begin{equation} \label{diag:conjugatechannels} \vcenter{\xymatrix{ P\ar[r]^-{c} & X\ar[r]^-{d} & O }} \qquad\qquad\mbox{or, as diagram,}\qquad\qquad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[arrow box] (c) at (0.0,0.5) {$c$}; \node[arrow box] (d) at (0.0,1.2) {$d$}; \coordinate (P) at (0.0,0.0); \coordinate (Y) at (0.0,1.7); \draw (P) to (c); \draw (c) to (d); \draw (d) to (Y); \end{tikzpicture}}} \end{equation} \noindent Such a pair of composable channels may be seen as a 2-stage hierarchical Bayesian model. In that context the parameters $P$ are sometimes called `hyperparameters', see \textit{e.g.}~\cite{BernardoS00}. There, esp.\ in Defn~5.6 of conjugate priorship one can also distinguish two channels, written as $p(\theta\mid\tau)$ and $p(x\mid\theta)$, corresponding respectively to our channels $c$ and $d$. The $\tau$ form the hyperparameters. In this setting we come to our main definition that formulates the notion of conjugate prior in an abstract manner, avoiding classes of distributions. It contains the crucial equation that was missing in the informal description in Section~\ref{sec:ideas}. All our examples of (conjugate prior) channels are maps in the Kleisli category of the Giry monad, but the formulation applies more generally. In fact, abstraction purifies the situation and shows the essentials. The definition below speaks of `deterministic' channels, between brackets. This part will be explained later on, in the beginning of Section~\ref{sec:conjugatepriorinversion}. It can be ignored for now. \begin{definition} \label{def:conjugateprior} In the situation~\eqref{diag:conjugatechannels} we call channel $c$ a \emph{conjugate prior} to channel $d$ if there is a (deterministic) channel $h\colon P\times O \rightarrow P$ for which the following equation holds: \begin{equation} \label{eqn:conjugateprior} \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[arrow box] (c) at (0.0,-0.2) {$c$}; \node[copier] (copier) at (0,0.3) {}; \node[arrow box] (d) at (0.5,0.95) {$d$}; \coordinate (X) at (-0.5,1.5); \coordinate (Y) at (0.5,1.5); \draw (c) to (copier); \draw (c) to (0.0,-0.7); \draw (copier) to[out=150,in=-90] (X); \draw (copier) to[out=15,in=-90] (d); \draw (d) to (Y); \end{tikzpicture}}} \quad = \quad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[copier] (copier1) at (0,0.3) {}; \node[copier] (copier2) at (0.25,2.0) {}; \coordinate (X) at (-0.5,3.8); \coordinate (Y) at (0.5,3.8); \node[arrow box] (c1) at (0.25,0.8) {$c$}; \node[arrow box] (d) at (0.25,1.5) {$d$}; \node[arrow box] (h) at (-0.5,2.6) {$\;\;h\;\;$}; \node[arrow box] (c2) at (-0.5,3.3) {$c$}; \draw (copier1) to (0.0,0.0); \draw (copier1) to[out=150,in=-90] (h.240); \draw (copier1) to[out=30,in=-90] (c1); \draw (c1) to (d); \draw (d) to (copier2); \draw (copier2) to[out=165,in=-90] (h.305); \draw (h) to (c2); \draw (c2) to (X); \draw (copier2) to[out=30,in=-90] (Y); \end{tikzpicture}}} \end{equation} \noindent Equivalently, in equational form: \[ \begin{array}{rcl} \tuple{\idmap, d} \after c & = & ((c\after h)\otimes\idmap) \after \tuple{\idmap, \copier \after d \after c}. \end{array} \] \end{definition} The idea is that the map $h\colon P\times O \rightarrow P$ translates parameters, with an observation from $O$ as additional argument. Informally, one gets a posterior state $c(h(p,y))$ from the prior state $c(p)$, given the observation $y\in O$. The power of this `analytic' approach is that it involves simple re-computation of parameters, instead of more complicated updating of entire states. This will be illustrated in several standard examples below. The above Equation~\eqref{eqn:conjugateprior} is formulated in an abstract manner --- which is its main strength. We will derive an alternative formulation of Equation~\eqref{eqn:conjugateprior} for pdf-channels. It greatly simplifies the calculations in examples. \begin{lemma} \label{lem:conjugatepriorpdf} Consider composable channels $\smash{P \stackrel{c}{\rightarrow} X \stackrel{d}{\rightarrow} O}$, as in~\eqref{diag:conjugatechannels}, for the Giry monad $\Giry$, where $c\colon P\rightarrow \Giry(X)$ and $d\colon X \rightarrow \Giry(O)$ are given by pdf's $u\colon P\times X \rightarrow \R_{\geq 0}$ and $v \colon X\times O\rightarrow \R_{\geq 0}$, as pdf-channels $c = \int u$ and $d = \int v$. Let $c$ be conjugate prior to $d$, via a measurable function $h\colon P\times O \rightarrow P$. Equation~\eqref{eqn:conjugateprior} then amounts to, for an element $p\in P$ and for measurable subsets $M\subseteq X$ and $N\subseteq O$, \begin{equation} \label{eqn:conjugatepriorint} \begin{array}{rcl} \lefteqn{\displaystyle\int_{N} \int_{M} u(p,x) \cdot v(x,y) \intd x\intd y} \\[+0.8em] & = & \displaystyle \int_{N} \Big(\int u(p,x)\cdot v(x,y)\intd x\Big) \cdot \Big(\int_{M} u(h(p,y),x) \intd x\Big) \intd y. \end{array} \end{equation} \noindent In order to prove this equation, it suffices to prove that the two functions under the outer integral $\int_{N}$ are equal, that is, it suffices to prove for each $y\in O$, \begin{equation} \label{eqn:conjugatepriorfun} \begin{array}{rcl} \displaystyle\int_{M} u(p,x) \cdot v(x,y) \intd x & = & \displaystyle \Big(\int u(p,x)\cdot v(x,y)\intd x\Big) \cdot \Big(\int_{M} u(h(p,y),x) \intd x\Big). \end{array} \end{equation} This formulation will be used in the examples below. \end{lemma} \begin{myproof} We extensively use the equations for integration from Section~\ref{sec:Kleisli} and from the Appendix, in order to prove~\eqref{eqn:conjugatepriorint}. The left-hand-side of Equation~\eqref{eqn:conjugateprior} gives the left-hand-side of~\eqref{eqn:conjugatepriorint}: \[ \begin{array}{rcccl} \big(\tuple{\idmap, d} \after c\big)(p)(M\times N) & \smash{\stackrel{\eqref{eqn:continuouscomposition}}{=}} & \big(\tuple{\idmap, d} \gg c(p)\big)(M\times N) & \smash{\stackrel{\eqref{eqn:pdfgraphstatetransformation}}{=}} & \displaystyle \int_{N} \int_{M} u(p,x) \cdot v(x,y) \intd x \intd y. \end{array} \] \noindent Unravelling the right-hand-side of~\eqref{eqn:conjugateprior} is a bit more work: \[ \begin{array}{rcl} \lefteqn{\big((c \after h)\otimes\idmap) \after \tuple{\idmap, \copier \after d \after c}\big)(p)(M\times N)} \\ & \smash{\stackrel{\eqref{eqn:continuouscomposition}}{=}} & \displaystyle \int (c \after h)\otimes\idmap)(-)(M\times N) \intd \tuple{\idmap, \copier \after d \after c}(p) \\ & \smash{\stackrel{\eqref{eqn:graphequation}}{=}} & \displaystyle \int ((c \after h)\otimes\idmap)(-)(M\times N) \intd \big(\eta(p) \otimes (\copier \after d \after c)(p)\big) \\ & \smash{\stackrel{\eqref{eqn:productpdfintegration}}{=}} & \displaystyle \int \int ((c \after h)\otimes\idmap)(-,-)(M\times N) \intd \eta(p) \intd (\copier \after d \after c)(p)\big) \\ & = & \displaystyle \int ((c \after h)\otimes\idmap)(p,-)(M\times N) \intd \Giry(\copier)(d \gg c(p)) \\ & \smash{\stackrel{\eqref{eqn:invmeasure}}{=}} & \displaystyle \int ((c \after h)\otimes\idmap)(p,\copier(-))(M\times N) \intd (d \gg c(p)) \\ & \smash{\stackrel{(\ref{eqn:pdfintegration}, \ref{eqn:pdfseqcomposition})}{=}} & \displaystyle \int \Big(\int u(p,x)\cdot v(x,y)\intd x\Big) \cdot ((c \after h)\otimes\idmap)(p,y,y)(M\times N) \intd y \\ & = & \displaystyle \int \Big(\int u(p,x)\cdot v(x,y)\intd x\Big) \cdot c(h(p,y))(M) \cdot \indic{N}(y) \intd y \\ & = & \displaystyle \int_{N} \Big(\int u(p,x)\cdot v(x,y)\intd x\Big) \cdot \Big(\int_{M} u(h(p,y),x) \intd x\Big) \intd y. \end{array} \] \noindent By combining this outcome with the earlier one we get the desired equation~\eqref{eqn:conjugatepriorint}. \QED \end{myproof} One can reorganise Equation~\eqref{eqn:conjugatepriorfun} as a normalisation fraction: \begin{equation} \label{eqn:conjugatepriorfunnorm} \begin{array}{rcl} \displaystyle\int_{M} u(h(p,y),x) \intd x & = & \displaystyle \frac{\int_{M} u(p,x) \cdot v(x,y) \intd x} {\int u(p,x)\cdot v(x,y)\intd x}. \end{array} \end{equation} \noindent It now strongly resembles Equation~\eqref{eqn:continuousinversion} for Bayesian inversion. This connection will be established more generally in Theorem~\ref{thm:conjugateinversion}. Essentially, the above normalisation fraction~\eqref{eqn:conjugatepriorfunnorm} occurs in~\cite[Defn.~5.6]{BernardoS00}. Later, in Section~\ref{sec:logic} we will see that~\eqref{eqn:conjugatepriorfunnorm} can also be analysed in terms of updating a state with a random variable. We are now ready to review some standard examples. The first one describes the structure underlying the coin example in Section~\ref{sec:ideas}. \begin{example} \label{ex:betaflip} It is well-known that the beta distributions are conjugate prior to the Bernoulli `flip' likelihood function. We shall re-formulate this fact following the pattern of Definition~\ref{def:conjugateprior}, with two composable channels, as in~\eqref{diag:conjugatechannels}, namely: \[\xymatrix@C+1pc{ \NNO_{>0}\times\NNO_{>0}\ar[r]^-{\betachan} & [0,1]\ar[r]^-{\flipchan} & 2 \rlap{\qquad where $2=\{0,1\}$.} } \qquad \] \noindent The $\betachan$ channel is as in~\eqref{diag:beta}, but now restricted to the non-negative natural numbers $\NNO_{>0}$. We recall that the normalisation constant $B(\alpha,\beta)$ is $\int_{[0,1]} x^{\alpha-1}(1-x)^{\beta-1} \intd x$. The $\flipchan$ channel sends a probability $r\in[0,1]$ to the $\text{Bernoulli}(r)$ distribution, which can also be written as a discrete distribution $\flipchan(r) = r\ket{1} + (1-r)\ket{0}$. More formally, as a Kleisli map $[0,1]\rightarrow\Giry(2)$ it is, for a subset $N\subseteq 2$, \[ \begin{array}{rcccccl} \flipchan(r)(N) & = & \displaystyle \int_{N} r^{i}\cdot (1-r)^{1-i} \intd i & = & \displaystyle\sum_{i\in N} r^{i}\cdot (1-r)^{1-i} & = & \left\{{\renewcommand{\arraystretch}{1.0}\begin{array}{ll} 0 & \mbox{if } N = \emptyset \\ r & \mbox{if } N = \{1\} \\ 1-r \;\; & \mbox{if } N = \{0\} \\ 1 & \mbox{if } N = \{0,1\}. \end{array}}\right. \end{array} \] \noindent The $i$ in $\intd i$ refers here to the counting measure. In order to show that $\betachan$ is a conjugate prior of $\flipchan$ we have to produce a parameter translation function $h\colon \NNO_{>0}\times\NNO_{>0}\times 2 \rightarrow \NNO_{>0}\times\NNO_{>0}$. It is defined by distinguishing the elements in $2 = \{0,1\}$ \begin{equation} \label{eqn:betaflipfun} \begin{array}{rclcrcl} h(\alpha, \beta, 1) & = & (\alpha+1, \beta) & \qquad\mbox{and}\qquad & h(\alpha, \beta, 0) & = & (\alpha, \beta+1). \end{array} \end{equation} \noindent Thus, in one formula, $h(\alpha,\beta,i) = (\alpha+i, \beta+(1-i))$. We prove Equation~\eqref{eqn:conjugatepriorfun} for $c = \betachan = \int u$ and $d = \flipchan = \int v$. We start from its right-hand-side, for an arbitrary $i\in 2$, \[ \begin{array}{rcl} \lefteqn{\displaystyle \Big(\int u(\alpha,\beta,x)\cdot v(x,i)\intd x\Big) \cdot \Big(\int_{M} u(h(\alpha,\beta,i),x) \intd x\Big)} \\ & = & \displaystyle\Big(\int \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)} \cdot x^{i}\cdot (1-x)^{1-i} \intd x\Big) \cdot \Big(\int_{M} \frac{x^{\alpha+i-1}(1-x)^{\beta+(1-i)-1}} {B(\alpha+i,\beta+(1-i))} \intd x\Big) \\[+1em] & = & \displaystyle\Big(\frac{\int x^{\alpha+i-1}(1-x)^{\beta+(1-i)-1} \intd x} {B(\alpha,\beta)}\Big) \cdot \Big(\int_{M} \frac{x^{\alpha-1}(1-x)^{\beta-1}} {B(\alpha+i,\beta+(1-i))} \cdot x^{i} \cdot (1-x)^{1-i} \intd x\Big) \\[+1em] & = & \displaystyle\Big(\frac{B(\alpha+i,\beta+(1-i))}{B(\alpha,\beta)}\Big) \cdot \Big(\int_{M} \frac{x^{\alpha-1}(1-x)^{\beta-1}} {B(\alpha+i,\beta+(1-i))} \cdot x^{i} \cdot (1-x)^{1-i} \intd x\Big) \\[+1em] & = & \displaystyle \int_{M}\frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)} \cdot x^{i} \cdot (1-x)^{1-i} \intd x \\[+1em] & = & \displaystyle \int_{M} u(\alpha,\beta,x)\cdot v(x, i) \intd x. \end{array} \] \noindent The latter expression is the left-hand-side of~\eqref{eqn:conjugatepriorfun}. We see that the essence of the verification of the conjugate prior equation is the shifting of functions and normalisation factors. This is a general pattern. \auxproof{ We have to prove Equation~\eqref{eqn:conjugateprior}. Its left-hand-side is, as Kleisli map $\NNO_{>0}\times\NNO_{>0} \rightarrow \Giry([0,1]\times 2)$, applied to measurable subsets $M\subseteq [0,1]$ and $N\subseteq 2$, \[ \begin{array}{rcl} \big(\tuple{\idmap, \flipchan} \after \betachan\big)(\alpha,\beta)(M\times N) & = & \big(\tuple{\idmap, \flipchan} \gg \betachan(\alpha,\beta)\big)(M\times N) \\ & \smash{\stackrel{\eqref{eqn:pdfgraphstatetransformation}}{=}} & \displaystyle \int_{M\times N} \frac{x^{\alpha-1}(1-x)^{\beta-1}} {B(\alpha,\beta)} \cdot x^{i}\cdot (1-x)^{1-i} \intd (x,i) \\ & = & \displaystyle \int_{M\times N} \frac{x^{\alpha-1+i}(1-x)^{\beta-i}} {B(\alpha,\beta)} \intd (x,i). \end{array} \] \noindent With a bit more effort we show that the right-hand-side of~\eqref{eqn:conjugateprior} has the same outcome: \[ \begin{array}{rcl} \lefteqn{\big(((\betachan \after h)\otimes\idmap) \after \tuple{\idmap, \copychan \after \flipchan \after \betachan}\big) (\alpha, \beta)(M\times N)} \\ & \smash{\stackrel{\eqref{eqn:graphequation}}{=}} & \displaystyle \int ((\betachan \after h)\otimes\idmap)(-)(M\times N) \intd \big(\eta(\alpha,\beta) \otimes (\copychan \after \flipchan \after \betachan)(\alpha,\beta)\big) \\ & = & \displaystyle \int ((\betachan \after h)\otimes\idmap)(\alpha, \beta, -) (M\times N) \intd \Giry(\copychan)\big((\flipchan \after \betachan)(\alpha,\beta)\big) \\ & \smash{\stackrel{\eqref{??}}{=}} & \displaystyle \int ((\betachan \after h)\otimes\idmap)(\alpha, \beta, \copychan(-))(M\times N) \intd \big(\flipchan \after \betachan)(\alpha,\beta)\big) \\ & \smash{\stackrel{(\ref{eqn:pdfintegration}, \ref{eqn:pdfseqcomposition})}{=}} & \displaystyle \int \big(\int \frac{x^{\alpha-1+i}(1-x)^{\beta-i}} {B(\alpha,\beta)}\intd x\big) \cdot ((\betachan \after h)\otimes\idmap)(\alpha, \beta, \copychan(i))(M\times N) \intd i \\ & = & \displaystyle \int \frac{\int x^{\alpha-1+i}(1-x)^{\beta-i} \intd x} {B(\alpha,\beta)} \cdot \betachan(h(\alpha, \beta, i))(M) \cdot \indic{N}(i) \intd i \\ & = & \displaystyle \int_{N} \frac{B(\alpha+i, \beta+1-i)}{B(\alpha,\beta)} \cdot \betachan(\alpha+i, \beta+1-i)(M) \intd i \\ & = & \displaystyle \int_{N} \frac{B(\alpha+i, \beta+1-i)}{B(\alpha,\beta)} \cdot \big(\int_{M} \frac{x^{\alpha+i-1}\cdot (1-x)^{\beta-i}}{B(\alpha+i, \beta+1-i)} \intd x\big) \intd i \\ & = & \displaystyle \int_{M\times N} \frac{x^{\alpha-1+i}(1-x)^{\beta-i}} {B(\alpha,\beta)} \intd (x,i). \end{array} \] } \end{example} \begin{example} \label{ex:betabinom} In a similar way one verifies that the $\betachan$ channel is a conjugate prior to the binomial channel. For the latter we fix a natural number $n>0$, and consider the two channels: \[ \xymatrix@-1pc{ \NNO_{>0}\times\NNO_{>0}\ar[rr]^-{\betachan} & & [0,1]\ar[rrr]^-{\binomchan_{n}} & & & \{0,1,\ldots,n\} } \] \noindent The binomial channel $\binomchan_{n}$ is defined for $r \in [0,1]$ and $M\subseteq \{0,1,\ldots,n\}$ as: \[ \begin{array}{rcccl} \binomchan_{n}(r)(M) & = & {\displaystyle \int_{M}} \binom{n}{i}\cdot r^{i}\cdot (1-r)^{n-i} \intd i & = & {\displaystyle \sum_{i\in M}}\, \binom{n}{i}\cdot r^{i}\cdot (1-r)^{n-i}. \end{array} \] \noindent The conjugate prior property requires in this situation a parameter translation function $h\colon \NNO_{>0}\times\NNO_{>0} \times \{0,1,\ldots,n\}\rightarrow \NNO_{>0}\times\NNO_{>0}$, which is given by: \[ \begin{array}{rcl} h(\alpha, \beta, i) & = & (\alpha+i, \beta+n-i). \end{array} \] \noindent The proof of Equation~\eqref{eqn:conjugatepriorfun} is much like in Example~\ref{ex:betaflip}, with $1-i$ replaced by $n-i$, and an additional binomial term $\binom{n}{i}$ that is shifted from one integral to another. \begin{Auxproof} \[ \begin{array}{rcl} \lefteqn{\displaystyle\Big(\int \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)} \cdot \binom{n}{i}\cdot x^{i} \cdot (1-x)^{n-i} \intd x\Big) \cdot \Big(\int_{M} \frac{x^{\alpha+i-1}(1-x)^{\beta+(n-i)-1}} {B(\alpha+i,\beta+(n-i))} \intd x\Big)} \\[+1em] & = & \displaystyle\Big(\frac{\int x^{\alpha+i-1}(1-x)^{\beta+(n-i)-1} \intd x} {B(\alpha,\beta)}\Big) \cdot \binom{n}{i} \cdot \Big(\int_{M} \frac{x^{\alpha-1}(1-x)^{\beta-1}} {B(\alpha+i,\beta+(n-i))} \cdot x^{i} \cdot (1-x)^{n-i} \intd x\Big) \\ & = & \displaystyle\Big(\frac{B(\alpha+i,\beta+(n-i))}{B(\alpha,\beta)}\Big) \cdot \Big(\int_{M} \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha+i,\beta+(n-i))} \cdot \binom{n}{i}\cdot x^{i} \cdot (1-x)^{n-i} \intd x\Big) \\ & = & \displaystyle \int_{M}\frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)} \cdot \binom{n}{i} \cdot x^{i} \cdot (1-x)^{n-i} \intd x. \end{array} \] \end{Auxproof} \end{example} Here is another well-known conjugate prior relationship, namely between Dirichlet and `multinomial' distributions. The latter are simply called discrete distributions in the present context. \begin{example} \label{ex:dirmon} Here we shall identify a number $n\in\NNO$ with the $n$-element set $\{0,1,\ldots,n-1\}$. We then write $\Dst_{*}(n)$ for the set of $n$-tuples $(x_{0}, \ldots, x_{n-1})\in (\R_{>0})^{n}$ with $\sum_{i}x_{i} = 1$. For a fixed $n>0$, let $O = \{y_{0}, \ldots, y_{n-1}\}$ be a set of `observations'. We consider the following two channels. \[ \xymatrix@-1pc{ (\NNO_{>0})^{n}\ar[rr]^-{\dirchan_n} & & \Dst_{*}(n)\ar[rr]^-{\multchan} & & O } \] \noindent The multinomial channel is defined as $\multchan(x_{0}, \ldots, x_{n-1}) = x_{0}\ket{y_{0}} + \cdots + x_{n-1}\ket{y_{n-1}}$. It can be described as a pdf-channel, via the function $v(\vec{x},y) \coloneqq x_{i} \mbox{ if }y=y_{i}$. Then, for $N\subseteq O = \{y_{0}, \ldots, y_{n-1}\}$, \[ \begin{array}{rcccl} \multchan(\vec{x})(N) & = & \displaystyle\int_{N} v(\vec{x}, y) \intd y & = & \sum\set{x_{i}}{y_{i}\in N}. \end{array} \] The Dirichlet channel $\dirchan_n$ is more complicated: for an $n$-tuple $\vec{\alpha} = (\alpha_{0}, \ldots, \alpha_{n-1})$ it is given via pdf's $d_n$, in: \[ \begin{array}{rclcrcl} \dirchan_{n}(\vec{\alpha}) & = & \displaystyle\int d_{n}(\vec{\alpha}) & \qquad\mbox{where}\qquad & d_{n}(\vec{\alpha})(x_{0}, \ldots, x_{n-1}) & = & {\displaystyle\frac{\Gamma(\sum_{i}\alpha_{i})}{\prod_{i}\Gamma(\alpha_{i})}} \cdot \prod_{i} x_{i}^{\alpha_{i}-1}, \end{array} \] \noindent for $(x_{0}, \ldots, x_{n-1}) \in \Dst_{*}(n)$. The operation $\Gamma$ is the `Gamma' function, which is defined on natural numbers $k > 1$ as $\Gamma(k) = (k-1)!$. Hence $\Gamma$ can be defined recursively as $\Gamma(1) = 1$ and $\Gamma(k+1) = k\cdot\Gamma(k)$. The above fraction is a normalisation factor since one has $\frac{\prod_{i} \Gamma(\alpha_{i})}{\Gamma(\sum_{i}\alpha_{i})} = \int\prod_{i} x_{i}^{\alpha_{i}-1} \intd \vec{x}$, see \textit{e.g.}~\cite{Bishop06}. From this one can derive: $\int x_{i}\cdot d_{n}(\vec{\alpha})(\vec{x}) \intd \vec{x} = \frac{\alpha_{i}}{\sum_{j}\alpha_j}$. The parameter translation function $h\colon (\NNO_{>0})^{n} \times O \rightarrow (\NNO_{>0})^{n}$ is: \[ \begin{array}{rcl} h(\alpha_{0}, \ldots, \alpha_{n-1}, y) & = & (\alpha_{0}, \ldots, \alpha_{i}+1, \ldots, \alpha_{n-1}) \quad \mbox{if } y=y_{i}. \end{array} \] \noindent We check Equation~\eqref{eqn:conjugatepriorfun}, for $M\subseteq \Dst_{*}(n)$ and observation $y_{i}\in O$, \[ \begin{array}{rcl} \lefteqn{\displaystyle \Big(\int d_{n}(\vec{\alpha})(\vec{x})\cdot v(\vec{x},y_{i})\intd\vec{x}\Big) \cdot \Big(\int_{M} d_{n}(h(\vec{\alpha},y_{i}))(\vec{x}) \intd \vec{x}\Big)} \\[+0.4em] & = & \frac{\alpha_i}{\sum_{j}\alpha_{j}} \cdot {\displaystyle\int_{M}} \frac{\Gamma(1 + \sum_{j}\alpha_{j})} {\Gamma(\alpha_{i}+1)\cdot \prod_{j\neq i}\Gamma(\alpha_{j})} \cdot x_{i}^{\alpha_i} \cdot \prod_{j\neq i} x_{j}^{\alpha_{j}-1} \intd \vec{x} \\[+0.8em] & = & {\displaystyle\Big(\int_{M}} \frac{\Gamma(\sum_{j}\alpha_{j})} {\prod_{j}\Gamma(\alpha_{j})} \cdot x_{i} \cdot \prod_{j} x_{j}^{\alpha_{j}-1} \intd \vec{x} \\[+0.8em] & = & \displaystyle\int_{M} d_{n}(\vec{\alpha})(\vec{x})\cdot v(\vec{x},y_{i})\intd\vec{x}. \end{array} \] \end{example} We include one more example, illustrating that normal channels are conjugate priors to themselves. This fact is also well-known. The point is to illustrate once again how that works in the current setting. \begin{example} \label{ex:normnorm} Consider the following two normal channels. \[ \xymatrix@-0.5pc{ \R\times\R_{>0}\ar[rr]^-{\normchan} & & \R\ar[rrr]^-{\normchan(-,\nu)} & & & \R_{>0} } \] \noindent The channel $\normchan$ is described explicitly in~\eqref{diag:normal}. Notice that we use it twice here, the second time with a fixed standard deviation $\nu$, for `noise'. This second channel is typically used for observation, like in Kalman filtering, for which a fixed noise level can be assumed. We claim that the first normal channel $\normchan$ is a conjugate prior to the second channel $\normchan(-,\nu)$, via the parameter translation function $h\colon \R\times\R_{>0}\times\R_{>0}\rightarrow \R\times\R_{>0}$ given by: \[ \begin{array}{rcl} h(\mu, \sigma, y) & = & \displaystyle (\, \frac{\mu\cdot\nu^{2} + y\cdot\sigma^{2}}{\nu^{2}+\sigma^{2}}, \frac{\nu\cdot\sigma}{\sqrt{\nu^{2}+\sigma^{2}}} \,) \end{array} \] \noindent We prove Equation~\eqref{eqn:conjugatepriorfun}, again starting from the right. \[ \begin{array}{rcl} \lefteqn{\Big({\displaystyle \int} \frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{(x-\mu)^{2}}{2\sigma^{2}}} \cdot \frac{1}{\sqrt{2\pi}\nu}e^{-\frac{(y-x)^{2}}{2\nu^{2}}}\intd x\Big) \cdot \Big({\displaystyle \int_{M}} \frac{\sqrt{\nu^{2}+\sigma^{2}}}{\sqrt{2\pi}\nu\sigma} e^{-\frac{(\nu^{2}+\sigma^{2}) (x-\frac{\mu\cdot\nu^{2} + y\cdot\sigma^{2}} {\nu^{2}+\sigma^{2}})^{2}}{2\nu^{2}\sigma^{2}}}\intd x\Big)} \\[+1em] & \smash{\stackrel{(*)}{=}} & \Big({\displaystyle \int}\frac{1}{\sqrt{2\pi}\sigma\nu} e^{-\frac{\nu^{2}(x-\mu)^{2} + \sigma^{2}(y-x)^{2}}{2\sigma^{2}\nu^{2}}}\intd x\Big) \\ & & \qquad \cdot\; \Big({\displaystyle \int_{M}} \frac{\sqrt{\nu^{2}+\sigma^{2}}}{\sqrt{2\pi}\nu\sigma} e^{-\frac{\nu^{2}(x-\mu)^{2} + \sigma^{2}(y-x)^{2} - \nu^{2}\mu^{2} - \sigma^{2}y^{2} + \frac{(\mu\cdot\nu^{2} + y\cdot\sigma^{2})^{2}} {\nu^{2}+\sigma^{2}}}{2\nu^{2}\sigma^{2}}}\intd x\Big) \\ & = & \Big({\displaystyle \int}\frac{\sqrt{\nu^{2}+\sigma^{2}}}{\sqrt{2\pi}\nu\sigma} e^{-\frac{\nu^{2}(x-\mu)^{2} + \sigma^{2}(y-x)^{2} - \nu^{2}\mu^{2} - \sigma^{2}y^{2} + \frac{(\mu\cdot\nu^{2} + y\cdot\sigma^{2})^{2}} {\nu^{2}+\sigma^{2}}}{2\sigma^{2}\nu^{2}}}\intd x\Big) \\ & & \qquad \cdot\; \Big({\displaystyle \int_{M}} \frac{1}{\sqrt{2\pi}\sigma\nu} e^{-\frac{\nu^{2}(x-\mu)^{2} + \sigma^{2}(y-x)^{2}}{2\nu^{2}\sigma^{2}}}\intd x\Big) \\ & \smash{\stackrel{(*)}{=}} & \Big({\displaystyle \int} \frac{\sqrt{\nu^{2}+\sigma^{2}}}{\sqrt{2\pi}\nu\sigma} e^{-\frac{(\nu^{2}+\sigma^{2}) (x-\frac{\mu\cdot\nu^{2} + y\cdot\sigma^{2}} {\nu^{2}+\sigma^{2}})^{2}}{2\nu^{2}\sigma^{2}}}\intd x\Big) \cdot \Big({\displaystyle \int_{M}} \frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{(x-\mu)^{2}}{2\sigma^{2}}} \cdot \frac{1}{\sqrt{2\pi}\nu}e^{-\frac{(y-x)^{2}}{2\nu^{2}}}\intd x\Big) \\[+1em] & = & {\displaystyle \int_{M}} \frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{(x-\mu)^{2}}{2\sigma^{2}}} \cdot \frac{1}{\sqrt{2\pi}\nu}e^{-\frac{(y-x)^{2}}{2\nu^{2}}}\intd x. \end{array} \] \noindent The last equation holds because the first integral in the previous line equals one, since, in general, the integral over a pdf is one. The two marked equations $\smash{\stackrel{(*)}{=}}$ are justified by: \[ \begin{array}{rcl} \lefteqn{(\nu^{2}+\sigma^{2})\big(x-\frac{\mu\cdot\nu^{2} + y\cdot\sigma^{2}}{\nu^{2}+\sigma^{2}}\big)^{2}} \\ & = & (\nu^{2}+\sigma^{2})x^{2}- 2(\mu\cdot\nu^{2} + y\cdot\sigma^{2})x + \frac{(\mu\cdot\nu^{2} + y\cdot\sigma^{2})^{2}} {\nu^{2}+\sigma^{2}} \\ & = & \nu^{2}(x^{2}-2\mu x + \mu^{2}) + \sigma^{2}(y^{2}-2yx + x^{2}) - \nu^{2}\mu^{2} - \sigma^{2}y^{2} + \frac{(\mu\cdot\nu^{2} + y\cdot\sigma^{2})^{2}} {\nu^{2}+\sigma^{2}} \\ & = & \nu^{2}(x-\mu)^{2} + \sigma^{2}(y-x)^{2} - \nu^{2}\mu^{2} - \sigma^{2}y^{2} + \frac{(\mu\cdot\nu^{2} + y\cdot\sigma^{2})^{2}} {\nu^{2}+\sigma^{2}} \end{array} \] \end{example} \section{Conjugate priors form Bayesian inversions}\label{sec:conjugatepriorinversion} This section connects the main two notions of this paper, by showing that conjugate priors give rise to Bayesian inversion. The argument is a very simple example of diagrammatic reasoning. Before we come to it, we have to clarify an issue that was left open earlier, regarding `deterministic' channels, see Definition~\ref{def:conjugateprior}. \begin{definition} \label{def:deterministic} A channel $c$ is called \emph{deterministic} if it commutes with copiers, that is, if it satisfies the equation on the left below. \[ \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[copier] (c) at (0,0.6) {}; \node[arrow box] (f) at (0,0.1) {$c$}; \draw (c) to[out=15,in=-90] (0.4,1.0); \draw (c) to[out=165,in=-90] (-0.4,1.0); \draw (f) to (c); \draw (f) to (0,-0.4); \end{tikzpicture}}} \quad=\quad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[copier] (c) at (0,0.0) {}; \node[arrow box] (f1) at (-0.4,0.6) {$c$}; \node[arrow box] (f2) at (0.4,0.6) {$c$}; \draw (c) to[out=165,in=-90] (f1); \draw (c) to[out=15,in=-90] (f2); \draw (f1) to (-0.4,1.1); \draw (f2) to (0.4,1.1); \draw (c) to (0,-0.3); \end{tikzpicture}}} \hspace*{10em} \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[copier] (c) at (0,0.4) {}; \node[state] (omega) at (0,0.1) {$\omega$}; \draw (c) to[out=15,in=-90] (0.4,0.8); \draw (c) to[out=165,in=-90] (-0.4,0.8); \draw (omega) to (c); \end{tikzpicture}}} \quad=\quad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (omega1) at (-0.5,0.0) {$\omega$}; \node[state] (omega2) at (0.5,0.0) {$\omega$}; \draw (omega1) to (-0.5,0.4); \draw (omega2) to (0.5,0.4); \end{tikzpicture}}} \] \noindent As a special case, a state $\omega$ is called deterministic if it satisfies the equation on the right, above. \end{definition} The state description is a special case of the channel description since a state on $X$ is a channel $1\rightarrow X$ and copying on the trivial (final) object $1$ does nothing, up to isomorphism. Few channels (or states) are deterministic. In deterministic and continuous computation, the ordinary functions $f\colon X \rightarrow Y$ are deterministic, when considered as a channel $\eta \after f$. We check this explicitly for point states, since this is what we need later on. \begin{example} \label{ex:deterministicstate} Let $x$ be an element of a measurable space $X$. The associated point state $\eta(x) \in \Giry(X)$ is deterministic, where $\eta(x)(M) = \indic{M}(x)$. We check the equation on the right in Definition~\ref{def:deterministic}: \[ \begin{array}{rcl} \big(\copier \after \eta(x)\big)(M\times N) & = & \eta(x,x)(M\times N) \hspace*{\arraycolsep}=\hspace*{\arraycolsep} \indic{M\times N}(x,x) \hspace*{\arraycolsep}=\hspace*{\arraycolsep} \indic{M}(x) \cdot \indic{N}(x) \\ & = & \eta(x)(M)\cdot \eta(x)(N) \hspace*{\arraycolsep}=\hspace*{\arraycolsep} \big(\eta(x)\otimes\eta(x)\big)(M\times N). \end{array} \] \end{example} We now come to the main result. \begin{theorem} \label{thm:conjugateinversion} Let $\smash{P \stackrel{c}{\rightarrow} X \stackrel{d}{\rightarrow} O}$ be channels, where $c$ is conjugate prior to $d$, say via $h\colon P \times O \rightarrow P$. Then for each deterministic (copyable) state $p$, the map $c \after h(p, -) \colon O \rightarrow X$ is a Bayesian inversion of $d$, wrt.\ the transformed state $c \gg p$. \end{theorem} \begin{myproof} We have to prove Equation~\eqref{diag:inversion}, for channel $d$ and state $c \gg p$, with the channel $c \after h(p, -)$ playing the role of Bayesian inversion $d^{\dag}_{c \gg p}$. This is easiest to see graphically, using that the state $p$ is deterministic and thus commutes with copiers $\copier$, see the equation on the right in Definition~\ref{def:deterministic}. \[ \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (p) at (0,-0.7) {$p$}; \node[arrow box] (c) at (0.0,-0.2) {$c$}; \node[copier] (copier) at (0,0.3) {}; \node[arrow box] (d) at (0.5,0.95) {$d$}; \coordinate (X) at (-0.5,1.5); \coordinate (Y) at (0.5,1.5); \draw (c) to (copier); \draw (c) to (p); \draw (copier) to[out=150,in=-90] (X); \draw (copier) to[out=15,in=-90] (d); \draw (d) to (Y); \end{tikzpicture}}} \quad \smash{\stackrel{\eqref{eqn:conjugateprior}}{=}} \quad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (p) at (0,0.0) {$p$}; \node[copier] (copier1) at (0,0.3) {}; \node[copier] (copier2) at (0.25,2.0) {}; \coordinate (X) at (-0.5,3.8); \coordinate (Y) at (0.5,3.8); \node[arrow box] (c1) at (0.25,0.8) {$c$}; \node[arrow box] (d) at (0.25,1.5) {$d$}; \node[arrow box] (h) at (-0.5,2.6) {$\;\;h\;\;$}; \node[arrow box] (c2) at (-0.5,3.3) {$c$}; \draw (copier1) to (p); \draw (copier1) to[out=150,in=-90] (h.240); \draw (copier1) to[out=30,in=-90] (c1); \draw (c1) to (d); \draw (d) to (copier2); \draw (copier2) to[out=165,in=-90] (h.305); \draw (h) to (c2); \draw (c2) to (X); \draw (copier2) to[out=30,in=-90] (Y); \end{tikzpicture}}} \quad = \quad \vcenter{\hbox{\begin{tikzpicture}[font=\small] \node[state] (p1) at (-0.65,2.1) {$p$}; \node[state] (p2) at (0.25,-0.15) {$p$}; \node[copier] (copier2) at (0.25,1.5) {}; \coordinate (X) at (-0.5,3.8); \coordinate (Y) at (0.5,3.8); \coordinate (Z) at (0.0,2.0); \node[arrow box] (c1) at (0.25,0.3) {$c$}; \node[arrow box] (d) at (0.25,1.0) {$d$}; \node[arrow box] (h) at (-0.5,2.6) {$\;\;h\;\;$}; \node[arrow box] (c2) at (-0.5,3.3) {$c$}; \draw (p1) to[out=90,in=-90] (h.240); \draw (p2) to[out=90,in=-90] (c1); \draw (c1) to (d); \draw (d) to (copier2); \draw (copier2) to[out=165,in=-90] (Z); \draw (Z) to[out=90,in=-90] (h.305); \draw (h) to (c2); \draw (c2) to (X); \draw (copier2) to[out=30,in=-90] (Y); \end{tikzpicture}}} \] \noindent This is it. \QED \end{myproof} When we specialise to Giry-channels we get an `if-and-only-if' statement, since there we can reason elementwise. \begin{corollary} \label{cor:Giry} Let $\smash{P\xrightarrow{c} X\xrightarrow{d} O}$ be two channels in $\Kl(\Giry)$, and let $h \colon P\times O \rightarrow P$ be a measurable function. The following two points are equivalent: \begin{itemize} \item[(i)] $c$ is a conjugate prior to $d$, via $h$; \item[(ii)] $c(h(p,-)) \colon O \rightarrow \Giry(X)$ is a Bayesian inverse for channel $d$ with state $c(p)$, \textit{i.e.}~is $d^{\dag}_{c(p)}$, for each parameter $p\in P$. \QED \end{itemize} \end{corollary} \section{A logical perspective on conjugate priors}\label{sec:logic} This section takes a logically oriented, look at conjugate priors, describing them in terms of updates of a prior state with a random variable (or predicate). This new perspective is interesting for two reasons: \begin{itemize} \item it formalises the intuition behind conjugate priors in a precise manner, see \textit{e.g.}\ Equations~\eqref{eqn:betaflipupdateeqn} and~\eqref{eqn:betabinomupdateeqn} below, where the characteristic closure property for a class of distributions is expressed via occurrences of these distributions on both sides of an equation; \item it will be useful in the next section to capture multiple observations via an update with a conjunction of multiple random variables. \end{itemize} \noindent But first we need to introduce some new terminology. We shall do so separately for discrete and continuous probability, although both can be described as instances of the same category theoretic notions, using effectus theory~\cite{Jacobs15d,Jacobs17a}. \subsection{Discrete updating} A \emph{random variable} on a set $X$ is a function $r\colon X \rightarrow \R$. It is a called a \emph{predicate} if it restricts to $X\rightarrow [0,1]$. Simple examples of predicates are indicator functions $\indic{E} \colon X \rightarrow [0,1]$, for a subset/event $E\subseteq X$, given by $\indic{E}(x) = 1$ if $x\in E$ and $\indic{E}(x) = 0$ if $x\not\in E$. Indicator functions $\indic{\{x\}} \colon X \rightarrow [0,1]$ for a singleton subset are sometimes called point predicates. For two random variables $r,s\colon X \rightarrow \R$ we write $r\andthen s\colon X \rightarrow \R$ for the new variable obtained by pointwise multiplication: $(r\andthen s)(x) = r(x) \cdot s(x)$. For a random variable $r$ and a discrete probability distribution (or state) $\omega\in\Dst(X)$ we define the \emph{validity} $\omega\models r$ as the expected value: \begin{equation} \label{eqn:discvalidity} \begin{array}{rcl} \omega\models r & \;\coloneqq\; & \displaystyle\sum_{x\in X} \omega(x)\cdot r(x). \end{array} \end{equation} \noindent Notice that this is a finite sum, since by definition the support of $\omega$ is finite. If we have a channel $c\colon X \rightarrow \Dst(Y)$ and a random variable $r\colon Y\rightarrow \R$ on its codomain $Y$, then we can transform it --- or pull it back --- into a random variable on its domain $X$. We write this pulled back random variable as $c \ll r \colon X \rightarrow \R$. It is defined as: \begin{equation} \label{eqn:discrandvartransform} \begin{array}{rcccl} \big(c \ll r\big)(x) & \coloneqq & c(x) \models r & = & \displaystyle\sum_{y\in Y} c(x)(y)\cdot r(y). \end{array} \end{equation} \noindent This operation $\ll$ interacts nicely with composition $\after$ of channels, in the sense that $(d\after c) \ll r = c \ll (d \ll r)$. Moreover, the validity $\omega \models c \ll r$ is the same as the validity $c\gg \omega\models r$, where $\gg$ is state transformation, see~\eqref{eqn:discstatransf}. If a validity $\omega\models r$ is non-zero, then we can define the \emph{updated} or \emph{conditioned} state $\omega|_{r} \in \Dst(X)$ via: \begin{equation} \label{eqn:discconditioning} \begin{array}{rclcrcl} \big(\omega|_{r}\big)(x) & \coloneqq & \displaystyle\frac{\omega(x)\cdot r(x)}{\omega\models r} & \qquad\mbox{that is}\qquad & \omega|_{r} & = & \displaystyle\sum_{x\in X} \frac{\omega(x)\cdot r(x)}{\omega\models r}\bigket{x}. \end{array} \end{equation} \noindent The first formulation describes the updated distribution $\omega|_{r}$ as a probability mass function, whereas the second one uses a formal convex sum. It is not hard to see that successive updates commute and can be reduced to a single update via $\andthen$, as in: \begin{equation} \label{eqn:discconditioningand} \begin{array}{rcccccl} \big(\omega|_{r}\big)|_{s} & = & \omega|_{r\andthen s} & = & \omega|_{s\andthen r} & = & \big(\omega|_{s}\big)|_{r}. \end{array} \end{equation} \noindent One can multiply a random variable $r\colon X \rightarrow \R$ with a scalar $a\in\R$, pointwise, giving a new random variable $a\cdot r \colon X \rightarrow \R$. When $a\neq 0$ it disappears from updating: \begin{equation} \label{eqn:discconditioningscal} \begin{array}{rcl} \omega|_{a\cdot r} & = & \omega|_{r}. \end{array} \end{equation} \begin{proposition} \label{prop:conjpriorpointupdate} Assume that composable channels $\smash{P \stackrel{c}{\rightarrow} X \stackrel{d}{\rightarrow} O}$ for the discrete distribution monad $\Dst$ are given, where $c$ is conjugate prior to $d$, say via $h\colon P \times O \rightarrow P$. The distribution for the updated parameter $h(p,y)$ is then an update of the distribution for the original parameter $p$, with the pulled-back point predicate for the observation $y$, as in: \[ \begin{array}{rcl} c\big(h(p,y)\big) & = & c(p)\big|_{d \ll \indic{\{y\}}}. \end{array} \] \end{proposition} \begin{myproof} We first notice that the pulled-back singleton predicate $d \ll \indic{\{y\}} \colon X \rightarrow \R$ is: \[ \begin{array}{rcccl} (d \ll \indic{\{y\}})(x) & \smash{\stackrel{\eqref{eqn:discrandvartransform}}{=}} & \displaystyle\sum_{z\in Y} d(x)(z) \cdot \indic{\{y\}}(z) & = & d(x)(y). \end{array} \] \noindent Theorem~\ref{thm:conjugateinversion} tells us that $c\big(h(p,y)\big)$ is obtained via the Bayesian inversion of $d$, so that: \[ \begin{array}[b]{rcl} c\big(h(p,y)\big)(x) & = & d^{\dag}_{c(p)}(y)(x) \\[+0.4em] & \smash{\stackrel{\eqref{eqn:discreteinversion}}{=}} & \displaystyle\frac{c(p)(x)\cdot d(x)(y)}{\sum_{z} c(p)(z)\cdot d(z)(y)} \\[+1em] & = & \displaystyle\frac{c(p)(x)\cdot (d \ll \indic{\{y\}})(x)} {\sum_{z} c(p)(z)\cdot (d \ll \indic{\{y\}})(z)} \qquad \mbox{as just noted} \\[+1em] & = & \displaystyle\frac{c(p)(x)\cdot (d \ll \indic{\{y\}})(x)} {c(p) \models d \ll \indic{\{y\}}} \\ & \smash{\stackrel{\eqref{eqn:discconditioning}}{=}} & c(p)\big|_{d \ll \indic{\{y\}}}. \end{array} \eqno{\QEDbox} \] \end{myproof} In fact, what we are using here is that the Bayesian inversion $c^{\dag}_{\omega}$ defined in~\eqref{eqn:discreteinversion} is an update: $c^{\dag}_{\omega} = \omega|_{c \ll \indic{\{y\}}}$. \subsection{Continuous updating} We now present the analogous story for continuous probability. A \emph{random variable} on a measurable space $X$ is a measurable function $X\rightarrow \R$. It is called a \emph{predicate} if it restricts to $X\rightarrow [0,1]$. These random variables (and predicates) are closed under $\andthen$ and scalar multiplication, defined via pointwise multiplication. In the continuous case one typically has no point predicates. Given a measure/state $\omega\in\Giry(X)$ and a random variable $r\colon X \rightarrow \R$ we define the validity $\omega\models r$ again as expected value: \begin{equation} \label{eqn:contvalidity} \begin{array}{rcl} \omega\models r & \;\coloneqq\; & \displaystyle\int r \intd\omega. \end{array} \end{equation} \noindent This allows us to define transformation of a random variable, backwards along a channel: for a channel $c\colon X\rightarrow \Giry(Y)$ and a random variable $r\colon Y\rightarrow \R$ we write $c \ll r \colon X \rightarrow \R$ for the pulled-back random variable defined by: \begin{equation} \label{eqn:contrandvartransform} \begin{array}{rcccl} \big(c \ll r\big)(x) & \coloneqq & c(x) \models r & = & \displaystyle\int r \intd\, c(x). \end{array} \end{equation} \noindent The update $\omega|_{r}\in\Giry(X)$ of a state $\omega\in\Giry(X)$ with a random variable $r\colon X \rightarrow \R$ is defined on a measurable subset $M\subseteq X$ as: \begin{equation} \label{eqn:contconditioning} \begin{array}{rcccccl} \big(\omega|_{r}\big)(M) & \coloneqq & \displaystyle\frac{\int_{M} r \intd\omega}{\int r \intd\omega} & = & \displaystyle\frac{\int_{M} r \intd\omega}{\omega\models r} & = & \displaystyle\int_{M}\frac{r}{\omega\models r} \intd\omega. \end{array} \end{equation} \noindent If $\omega = \int f$ for a pdf $f$, this becomes: \begin{equation} \label{eqn:contconditioningpdf} \begin{array}{rcccccl} \big(\omega|_{r}\big)(M) & \coloneqq & \displaystyle\frac{\int_{M} f(x) \cdot r(x) \intd x} {\int f(x) \cdot r(x) \intd x} & = & \displaystyle\frac{\int_{M} f(x) \cdot r(x) \intd x}{\omega\models r} & = & \displaystyle\int_{M} \frac{f(x) \cdot r(x)}{\omega\models r} \intd x. \end{array} \end{equation} \noindent The latter formulation shows that the pdf of $\omega|_{r}$ is the function $x \mapsto \frac{f(x) \cdot r(x)}{\omega\models r}$. Updating in the continuous case also satisfies the multiple-update and scalar properties~\eqref{eqn:discconditioningand} and ~\eqref{eqn:discconditioningscal}. Again we redescribe conjugate priors in terms of updating. \begin{proposition} \label{prop:contpriorpointupdate} Let $\smash{P \stackrel{c}{\rightarrow} X \stackrel{d}{\rightarrow} O}$ be channels for the Giry monad $\Giry$, where $c$ and $d$ are pdf-channels $c = \int u$ and $d = \int v$, for $u\colon P\times X \rightarrow \R_{\geq 0}$ and $v\colon X\times O \rightarrow \R_{\geq 0}$. Assume that $c$ be conjugate prior to $d$ via $h\colon P \times O \rightarrow P$. Then: \[ \begin{array}{rcl} c\big(h(p,y)\big) & = & c(p)\big|_{v(-,y)}, \end{array} \] \noindent where $v(-,y) \colon O \rightarrow \R$ is used as random variable on $O$. \end{proposition} \begin{myproof} Theorem~\ref{thm:conjugateinversion} gives the first step in: \[ \begin{array}[b]{rcl} c\big(h(p,y)\big)(M) & = & d^{\dag}_{c(p)}(y)(M) \\[+0.4em] & \smash{\stackrel{\eqref{eqn:continuousinversion}}{=}} & \displaystyle\frac{\int_{M} u(p,x)\cdot v(x,y) \intd x} {\int u(p,x) \cdot v(x,y) \intd x} \\[+1em] & = & \displaystyle\frac{\int_{M} u(p,x)\cdot v(x,y) \intd x} {c(p) \models v(-,y)} \\ & \smash{\stackrel{\eqref{eqn:contconditioningpdf}}{=}} & c(p)\big|_{v(-,y)}(M). \end{array} \eqno{\QEDbox} \] \end{myproof} The previous two propositions deal with two \emph{discrete} channels $c,d$ (for $\Dst$) or with two \emph{continuous} channels (for $\Giry$). But the update approach also works for mixed channels, technically because $\Dst$ is a submonad of $\Giry$. We shall not elaborate these details but give illustrations instead. \begin{example} \label{ex:betaflipupdate} We shall have another look at the $\betachan - \flipchan$ conjugate prior situation from Example~\ref{ex:betaflip}. We claim that the essence of these channels being conjugate prior, via the parameter translation function~\eqref{eqn:betaflipfun}, can be expressed via the following two state update equations: \begin{equation} \label{eqn:betaflipupdateeqn} \begin{array}{rcl} \betachan(\alpha,\beta)\big|_{\flipchan \ll \indic{\{1\}}} & = & \betachan(\alpha+1,\beta) \\ \betachan(\alpha,\beta)\big|_{\flipchan \ll \indic{\{0\}}} & = & \betachan(\alpha,\beta+1). \end{array} \end{equation} \noindent These equations follow from what we have proven above. But we choose to re-prove them here in order to illustrate how updating works concretely. First note that for a parameter $x\in [0,1]$ we have predicate values $\big(\flipchan \ll \indic{\{1\}}\big)(x) = x$ and $\big(\flipchan \ll \indic{\{0\}}\big)(x) = 1-x$. Then: \[ \begin{array}{rcl} \betachan(\alpha,\beta) \models \flipchan \ll \indic{\{1\}} \hspace*{\arraycolsep}\smash{\stackrel{\eqref{eqn:beta}}{=}}\hspace*{\arraycolsep} \displaystyle\int x\cdot \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)}\intd x & = & \displaystyle\frac{\int x^{\alpha}(1-x)^{\beta-1}\intd x}{B(\alpha,\beta)} \\[+0.8em] & = & \displaystyle\frac{B(\alpha+1,\beta)}{B(\alpha,\beta)}. \end{array} \] \noindent Thus, using~\eqref{eqn:contconditioningpdf}, we obtain the first equation in~\eqref{eqn:betaflipupdateeqn}: \[ \begin{array}{rcl} \betachan(\alpha,\beta)\big|_{\flipchan \ll \indic{\{1\}}}(M) & = & \displaystyle\frac{B(\alpha,\beta)}{B(\alpha+1,\beta)} \cdot \int_{M} x\cdot \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha,\beta)}\intd x \\[+0.8em] & = & \displaystyle\int_{M} \frac{x^{\alpha}(1-x)^{\beta-1}}{B(\alpha+1,\beta)}\intd x \\ & = & \betachan(\alpha+1,\beta)(M). \end{array} \] In a similar way we can capture the $\betachan - \binomchan$ conjugate priorship from Example~\ref{ex:betabinom} as update equation: \begin{equation} \label{eqn:betabinomupdateeqn} \begin{array}{rcl} \betachan(\alpha,\beta)\big|_{\binomchan_{n} \ll \indic{\{i\}}} & = & \betachan(\alpha+i, \beta+n-i). \end{array} \end{equation} \noindent This equation, and also~\eqref{eqn:betaflipupdateeqn}, hightlight the original ideal behind conjugate priors, expressed informally in many places in the literature as: we have a class of distributions --- $\betachan$ in this case --- which is closed under updates in a particular statistiscal model --- $\flipchan$ or $\binomchan$ in these cases. \end{example} These update formulations~\eqref{eqn:betabinomupdateeqn} and~\eqref{eqn:betaflipupdateeqn} may be useful when trying to find a parameter translation function: one can start calculating the state update on the left-hand-side, using formulas~\eqref{eqn:discconditioning} and~\eqref{eqn:contconditioning}, hoping that a distribution of the same form appears (but with different parameters). \section{Multiple updates}\label{sec:multiple} So far we have dealt with the situation where there is a single observation $y\in O$ that leads to an update of a prior distribution. In this final section we briefly look at how to handle multiple observations $y_{1}, \ldots, y_{m}$. This is what typically happens in practice; it will lead to the notion of \emph{sufficient statistic}. A good starting point is the $\betachan - \flipchan$ relationship from Example~\ref{ex:betaflip} and~\ref{ex:betaflipupdate}, especially in its snappy update form~\eqref{eqn:betaflipupdateeqn}. Suppose we have multiple head/tail observations $y_{1}, \ldots, y_{m} \in 2 = \{0,1\}$ which we wish to incorporate into a prior distribution $\betachan(\alpha,\beta)$. Following Equation~\eqref{eqn:betaflipupdateeqn} we use multiple updates, on the left below, which can be rewritten as a single update, on the right-hand-side of the equation via conjunction $\andthen$, using~\eqref{eqn:discconditioningand}: \[ \begin{array}{rcl} \lefteqn{\betachan(\alpha,\beta)\big|_{\flipchan \ll \indic{\{y_1\}}} \big|_{\flipchan \ll \indic{\{y_2\}}} \;\cdots\; \big|_{\flipchan \ll \indic{\{y_m\}}}} \\ & = & \betachan(\alpha,\beta)\big|_{(\flipchan \ll \indic{\{y_1\}}) \,\andthen\, (\flipchan \ll \indic{\{y_2\}}) \,\andthen\; \cdots \;\andthen\, (\flipchan \ll \indic{\{y_m\}})}. \end{array} \] \noindent The $m$-ary conjunction predicate in the latter expression amounts to $q(x) = x^{n_1}(1-x)^{n_0}$ where $n_{1} = \sum_{i} y_{i}$ is the number of $1$'s among the observation $y_i$ and $n_{0} = \sum_{i} (1-y_{i})$ is the number of $0$'s, see Example~\ref{ex:betaflipupdate}. Of course the outcome is $\betachan(\alpha+n_{1}, \beta+n_{0})$. The question that is relevant in this setting is: can a random variable $p(y_{1}, \ldots, y_{m})$ with many parameters somehow be simplified, like in $q$ above. This is where the notion of sufficient statistic arises, see \textit{e.g.}~\cite{Koopman36,Bishop06}. \begin{definition} \label{def:sufstat} Let $p\colon X\times O^{m} \rightarrow \R$ be a random variable, with $1+m$ inputs. A \emph{sufficient statistic} for $p$ is a triple of functions \[ \xymatrix{ O^{m}\ar[r]^-{s} & \R \qquad\quad O^{m}\ar[r]^-{t} & Z \qquad\quad X\times Z\ar[r]^-{q} & \R } \] \noindent so that $p$ can be written as: \begin{equation} \label{eqn:sufstat} \begin{array}{rcl} p(x, y_{1}, \ldots, y_{m}) & = & s(y_{1}, \ldots, y_{m}) \cdot q\big(x, t(y_{1}, \ldots, y_{m})\big). \end{array} \end{equation} \end{definition} In the above $\betachan$ example we would like to simplify the big conjunction random variable: \[ \begin{array}{rcl} p(x, y_{1}, \ldots, y_{m}) & = & \Big((\flipchan \ll \indic{\{y_1\}}) \,\andthen\; \cdots \;\andthen\, (\flipchan \ll \indic{\{y_m\}})\Big)(x). \end{array} \] \noindent We can take $Z = \NNO\times\NNO$ with $t(y_{1}, \ldots, y_{m}) = (n_{1}, n_{0})$, where $n_1$ and $n_{0}$ are the number of $1$'s and $0$'s in the $y_i$. Then $q(x, n, n') = x^{n}(1-x)^{n'}$. The function $s$ is trivial and sends everything to $1$. A sufficient statistic thus summarises, esp.\ via the function $t$, the essential aspects of a list of observations, in order to simplify the update. In the coin example, these essential aspects are the numbers of $1$'s and $0$'s (that is, of heads and tails). In these situations the conjunction predicate --- like $p$ above --- is usally called a \emph{likelihood}. The big advantage of writing a random variable $p$ in the form of~\eqref{eqn:sufstat} is that updating with $p$ can be simplified. Let $\omega$ be a distribution on $X$, either discrete or continuous. Then, writing $\vec{y} = (y_{1}, \ldots, y_{m})$ we get: \[ \begin{array}{rcccl} \omega|_{p(-,\vec{y})} & = & \omega|_{s(\vec{y})\cdot q(-,t(\vec{y}))} & = & \omega|_{q(-,t(\vec{y}))}. \end{array} \] \noindent The factor $s(\vec{y})$ drops out because it works like a scalar, see~\eqref{eqn:discconditioningscal}. We conclude this section with a standard example of a sufficient statistic (see \textit{e.g.}~\cite{Bishop06}), for a conjunction expression arising from multiple updates. \begin{example} \label{ex:normnormstat} Recall the $\normchan - \normchan$ conjugate priorship from Example~\ref{ex:normnorm}. The first channel there has the form $\normchan = \int u$, for $u(\mu,\sigma,x) = \nicefrac{1}{\sqrt{2\pi}\sigma}\cdot e^{-\nicefrac{(x-\mu)^{2}}{2\sigma^2}}$. The second channel is $\normchan(-,\nu) = \int v$, for a fixed `noise' factor $\nu$, where $v(x,y) = \nicefrac{1}{\sqrt{2\pi}\nu}\cdot e^{-\nicefrac{(y-x)^{2}}{2\nu^2}}$. Let's assume that we have observations $y_{1}, \ldots, y_{m} \in \R_{> 0}$ which we like to use to iteratively update the prior distribution $\normchan(\mu,\sigma)$. Following Proposition~\ref{prop:contpriorpointupdate} we can describe these updates as: \[ \begin{array}{rcl} \normchan(\mu,\sigma)\big|_{v(-,y_1)} \;\cdots\; \big|_{v(-,y_m)} & = & \normchan(\mu,\sigma)\big|_{v(-,y_1) \,\andthen \;\cdots\; \andthen\,v(-,y_m)}. \end{array} \] \noindent Thus we are interested in finding a sufficient statistics for the predicate: \[ \begin{array}{rcl} p(x,y_{1}, \ldots, y_{m}) & \coloneqq & \Big(v(-,y_1) \,\andthen \;\cdots\; \andthen\,v(-,y_m)\Big)(x) \\ & = & v(x,y_{1}) \cdot \;\cdots\; \cdot v(x,y_{m}) \\ & = & \frac{1}{\sqrt{2\pi}\nu}\cdot e^{-\frac{(y_{1}-x)^{2}}{2\nu^{2}}} \;\cdots\; \frac{1}{\sqrt{2\pi}\nu}\cdot e^{-\frac{(y_{m}-x)^{2}}{2\nu^{2}}} \\ & = & \frac{1}{(\sqrt{2\pi}\nu)^{m}}\cdot e^{-\frac{\sum_{i}(y_{i}-x)^{2}}{2\nu^2}} \\ & = & \frac{1}{(\sqrt{2\pi}\nu)^{m}}\cdot e^{-\frac{\sum_{i}y_{i}^{2}}{2\nu^2}} \cdot e^{\frac{2(\sum_{i}y_{i})x-mx^{2}}{2\nu^2}} \\ & = & s(y_{1}, \ldots, y_{m}) \cdot q\big(x, t(y_{1}, \ldots, y_{m})\big), \end{array} \] \noindent for functions $s,t,q$ given by: \[ \begin{array}{rclcrclcrcl} s(y_{1}, \ldots, y_{m}) & = & \frac{1}{(\sqrt{2\pi}\nu)^{m}}\cdot e^{-\frac{\sum_{i}y_{i}^{2}}{2\nu^2}} & \quad & t(y_{1}, \ldots, y_{m}) & = & \sum_{i}x_{i} & \quad & q(x, z) & = & e^{\frac{2zx-mx^{2}}{2\nu^2}}. \end{array} \] \end{example} \section{Conclusions}\label{sec:conclusions} This paper contains a novel view on conjugate priors, using the concept of channel in a systematic manner. It has introduced a precise definition for conjugate priorship, using a pair of composable channels $P\rightarrow X\rightarrow O$ and a parameter translation function $P\times O \rightarrow P$, satisfying a non-trivial equation, see Definition~\ref{def:conjugateprior}. It has been shown that this equation holds for several standard conjugate prior examples. There are many more examples, that have not been checked here. One can be confident that the same equation holds for those unchecked examples too, since it has been shown here that conjugate priors amount to Bayesian inversions. This inversion property is the essential characteristic for conjugate priors. It has been re-formulated in logical terms, so that the closure property of a class of priors under updating is highlighted. \appendix \section{Calculation laws for Giry-Kleisli maps with pdf's}\label{sec:calculation} We assume that for a probability distribution (state) $\omega\in\Giry(X)$ and a measurable function $f\colon X \rightarrow \R_{\geq 0}$ the integral $\int f \intd \omega \in [0,\infty]$ can be defined as a limit of integrals over simple functions that approximate $f$. We shall follow the description of~\cite{Jacobs13a}, to which we refer for details\footnote{In~\cite{Jacobs13a} integration $\int f\intd \omega$ is defined only for $[0,1]$-valued functions $f$, but that does not matter for the relevant equations, except that integrals may not exist for $\R_{\geq 0}$-valued functions (or have value $\infty$). These integrals are determined by their valued $\int \indic{M}\intd \omega = \omega(M)$ on indicator functions $\indic{M}$ for measurable subsets, via continuous and linear extensions, see also~\cite{JacobsW15a}.}. This integration satisfies the Fubini property, which can be formulated, for states $\omega\in\Giry(X)$, $\rho\in\Giry(Y)$ and measurable function $h\colon X\times Y \rightarrow \R_{\geq 0}$, as: \begin{equation} \label{eqn:productpdfintegration} \begin{array}{rcl} \displaystyle\int h \intd (\omega\otimes\rho) & = & \displaystyle\int \int h \intd \omega \intd\rho. \end{array} \end{equation} \noindent The product state $\omega\otimes\rho \in \Giry(X\times Y)$ is defined by $(\omega\otimes\rho)(M\times N) = \omega(M)\cdot\rho(N)$. \begin{Auxproof} It suffices to consider the special cases where $g = \indic{M}$ and $h = \indic{M\times N}$, in: \[ \begin{array}{rcl} \displaystyle \int \indic{M} \intd \omega & = & \omega(\indic{M}) \\ & = & \displaystyle \int_{M} f(x)\intd x \\ & = & \displaystyle\int f(x) \cdot \indic{M}(x) \intd x \\ & = & \displaystyle\int f(x) \cdot g(x) \intd x \\ \displaystyle \int \indic{M\times N} \intd (\omega\otimes\rho) & = & (\omega\otimes\rho)(M\times N) \\ & = & \omega(M) \cdot \rho(N) \\ & = & \big(\displaystyle\int_{M} f(x) \intd x\big)\cdot \big(\displaystyle \int \indic{N} \intd \rho\big) \\ & = & \displaystyle \int \big(\displaystyle\int_{M} f(x) \intd x\big)\cdot \indic{N} \intd \rho \\ & = & \displaystyle \int \int \indic{M}(x)\cdot f(x)\cdot \indic{N} \intd x \intd \rho \\ & = & \displaystyle \int \int \indic{M\times N}(x,-)\cdot f(x) \intd x \intd \rho \\ & = & \displaystyle \int \int h(x,-)\cdot f(x) \intd x \intd \rho. \end{array} \] \end{Auxproof} \subsection{States via pdf's}\label{subsec:stateviapdf} For a subset $X\subseteq \R$, a measurable function $f\colon X \rightarrow \R_{\geq 0}$ is called a probability density function (pdf) for a state $\omega\in\Giry(X)$ if $\omega(M) = \int_{M} f(x) \intd x$ for each measurable subset $M\subseteq X$. In that case we simply write $\omega = \int f(x) \intd x$, or even $\omega = \int f$. If $\omega$ is given by such a pdf $f$, integration with state $\omega$ can be described as: \begin{equation} \label{eqn:pdfintegration} \begin{array}{rcl} \displaystyle\int g \intd \omega & = & \displaystyle\int f(x)\cdot g(x) \intd x. \end{array} \end{equation} \subsection{Channels via pdf's}\label{subsec:channelviapdf} Let channel $c\colon X \rightarrow \Giry(Y)$ be given as $c = \int u$ by pdf $u\colon X\times Y \rightarrow \R_{\geq 0}$ as $c(x)(N) = \int_{N} u(x,y) \intd y$, for each $x\in X$ and measurable $N\subseteq Y$, like in~\eqref{eqn:channelfrompdf}. If $\omega = \int f$ is a state on $X$, then state transformation $c \gg \omega \in \Giry(Y)$ is given by: \begin{equation} \label{eqn:pdfstatetransformation} \hspace*{-0.8em}\begin{array}{rcl} (c \gg \omega)(N) \hspace*{\arraycolsep}\smash{\stackrel{\eqref{eqn:statetransformation}}{=}}\hspace*{\arraycolsep} \displaystyle \int c(-)(N) \intd \omega & \smash{\stackrel{\eqref{eqn:pdfintegration}}{=}} & \displaystyle \int f(x) \cdot c(x)(N) \intd x \\ & = & \displaystyle \int_{N} \int f(x) \cdot u(x,y) \intd x \intd y. \end{array} \end{equation} \noindent Hence the pdf of the transformed state $c \gg \omega$ is $y \mapsto \int f(x) \cdot u(x,y) \intd x$. Given a channel $d \colon Y \rightarrow \Giry(Z)$, say with $d = \int v$, then sequential channel composition $d \after c$ is given, for $x\in X$ and $K\subseteq Z$, by: \begin{equation} \label{eqn:pdfseqcomposition} \begin{array}{rcl} (d \after c)(x)(K) \hspace*{\arraycolsep}\smash{\stackrel{\eqref{eqn:continuouscomposition}}{=}}\hspace*{\arraycolsep} \displaystyle \int d(-)(K) \intd c(x) & \smash{\stackrel{\eqref{eqn:pdfintegration}}{=}} & \displaystyle \int u(x,y) \cdot d(y)(K) \intd y \\ & = & \displaystyle \int_{K} \int u(x,y) \cdot v(y,z) \intd y \intd z \end{array} \end{equation} \noindent We see that the pdf of the channel $d \after c$ is $(x,z) \mapsto \int u(x,y) \cdot v(y,z) \intd y$. For a channel $e = \int w \colon A \rightarrow \Giry(B)$ we get a parallel composition channel $c\otimes e \colon X\times A \rightarrow \Giry(Y\times B)$ given by: \begin{equation} \label{eqn:pdfparcomposition} \begin{array}{rcl} (c \otimes e)(x,a)(M\times N) & = & c(x)(M) \otimes e(a)(N) \\ & = & \displaystyle\big(\int_{M} u(x,y)\intd y\big)\cdot \big(\int_{N} w(a,b) \intd b\big) \\[0.8em] & = & \displaystyle \int_{M\times N} u(x,y)\cdot w(a,b) \intd (y,b). \end{array} \end{equation} \noindent Hence the pdf of the channel $c\otimes d$ is $(x,a,y,b) \mapsto u(x,y)\cdot w(a,b)$. \subsection{Graph channels and pdf's}\label{subsec:graphandpdf} For a channel $c\colon X \rightarrow \Giry(Y)$ we can form `graph' channels $\tuple{\idmap,c} = (\idmap\otimes c) \after \copier \colon X \rightarrow \Giry(X\times Y)$ and $\tuple{c,\idmap} = (c\otimes\idmap) \after \copier \colon X \rightarrow \Giry(Y\times X)$. For $x\in X$ we have: \begin{equation} \label{eqn:graphequation} \begin{array}{rclcrcl} \tuple{\idmap,c}(x) & = & \eta(x)\otimes c(x) & \qquad\mbox{and}\qquad & \tuple{c,\idmap}(x) & = & c(x) \otimes \eta(x). \end{array} \end{equation} \begin{Auxproof} \[ \begin{array}{rcl} \tuple{\idmap, c}(x)(M\times N) & = & \big(\st_{2} \after (\idmap\times c) \after \Delta\big)(x)(M\times N) \\ & = & \big(\dst \after (\eta\times c) \after \Delta\big)(x)(M\times N) \\ & = & \dst\big((\eta\times c)(x,x)\big)(M\times N) \\ & = & \dst\big(\eta(x), c(x)\big)(M\times N) \\ & = & \eta(x)(M) \cdot c(x)(N) \\ & = & \big(\eta(x) \otimes c(x)\big)(M\times N). \end{array} \] \end{Auxproof} \noindent If $c = \int u$ and $\omega = \int f$ is a state on $X$, then: \begin{equation} \label{eqn:pdfgraphstatetransformation} \begin{array}{rcl} (\tuple{\idmap,c} \gg \omega)(M\times N) & \smash{\stackrel{\eqref{eqn:pdfintegration}}{=}} & \displaystyle \int f(x) \cdot \tuple{\idmap,c}(x)(M\times N) \intd x \\ & \smash{\stackrel{\eqref{eqn:graphequation}}{=}} & \displaystyle \int f(x) \cdot \eta(x)(M) \cdot c(x)(N) \intd x \\ & = & \displaystyle \int_{N} \int_{M} f(x) \cdot u(x,y) \intd x \intd y. \end{array} \end{equation} \noindent We also consider the situation where $d\colon X\times Y \rightarrow \Giry(Z)$ is of the form $d = \int v$, with composition: \begin{equation} \label{eqn:pdfgraphcomposition} \begin{array}{rcl} \big(d \after \tuple{\idmap,c}\big)(x)(K) & \smash{\stackrel{\eqref{eqn:graphequation}}{=}} & \displaystyle \int d(-)(K) \intd (\eta(x)\otimes c(x)) \\ & \smash{\stackrel{\eqref{eqn:productpdfintegration}}{=}} & \displaystyle \int d(-)(K) \intd \eta(x) \intd c(x) \\ & = & \displaystyle \int d(x,-)(K) \intd c(x) \\ & \smash{\stackrel{\eqref{eqn:pdfintegration}}{=}} & \displaystyle \int u(x,y) \cdot d(x,y)(K) \intd y \\ & = & \displaystyle \int_{K} \int u(x,y) \cdot v(x,y,z) \intd y \intd z. \end{array} \end{equation} \noindent Hence the pdf of the channel $d \after \tuple{\idmap,c}$ is $(x,z) \mapsto \int u(x,y) \cdot v(x,y,z) \intd y$. \begin{thebibliography}{} \bibitem[Abramsky and Coecke, 2009]{AbramskyC09} Abramsky, S. and Coecke, B. (2009). \newblock A categorical semantics of quantum protocols. \newblock In Engesser, K., Gabbay, D.~M., and Lehmann, D., editors, {\em Handbook of Quantum Logic and Quantum Structures: Quantum Logic}, pages 261--323. North-Holland, Elsevier, Computer Science Press. \bibitem[Ackerman et~al., 2011]{AckermanFR11} Ackerman, N., Freer, C., and Roy, D. (2011). \newblock Noncomputable conditional distributions. \newblock In {\em Logic in Computer Science}. IEEE, Computer Science Press. \bibitem[Alpaydin, 2010]{Alpaydin10} Alpaydin, E. (2010). \newblock {\em Introduction to Machine Learning}. \newblock {MIT} Press, Cambridge, MA, $2^{\textrm{nd}}$ edition. \bibitem[Bernardo and Smith, 2000]{BernardoS00} Bernardo, J. and Smith, A. (2000). \newblock {\em Bayesian Theory}. \newblock John Wiley \& Sons. \bibitem[Bishop, 2006]{Bishop06} Bishop, C. (2006). \newblock {\em Pattern Recognition and Machine Learning}. \newblock Information Science and Statistics. Springer. \bibitem[Blackwell, 1951]{Blackwell51} Blackwell, D. (1951). \newblock Comparison of experiments. \newblock In {\em Proc. Sec. Berkeley Symp. on Math. Statistics and Probability}, pages 93--102. Springer/British Computer Society. \bibitem[Cho and Jacobs, 2017]{ChoJ17b} Cho, K. and Jacobs, B. (2017). \newblock The {EfProb} library for probabilistic calculations. \newblock In Bonchi, F. and K{\"o}nig, B., editors, {\em Conference on Algebra and Coalgebra in Computer Science (CALCO 2017)}, volume~72 of {\em LIPIcs}. Schloss Dagstuhl. \bibitem[Cho et~al., 2015]{ChoJWW15b} Cho, K., Jacobs, B., Westerbaan, A., and Westerbaan, B. (2015). \newblock An introduction to effectus theory. \newblock see \url{arxiv.org/abs/1512.05813}. \bibitem[Clerc et~al., 2017]{ClercDDG17} Clerc, F., Dahlqvist, F., Danos, V., and Garnier, I. (2017). \newblock Pointless learning. \newblock In Esparza, J. and Murawski, A., editors, {\em Foundations of Software Science and Computation Structures}, number 10203 in Lect. Notes Comp. Sci., pages 355--369. Springer, Berlin. \bibitem[Culbertson and Sturtz, 2014]{CulbertsonS14} Culbertson, J. and Sturtz, K. (2014). \newblock A categorical foundation for bayesian probability. \newblock {\em Appl. Categorical Struct.}, 22(4):647--662. \bibitem[Diaconis and Ylvisaker, 1979]{DiaconisY79} Diaconis, P. and Ylvisaker, D. (1979). \newblock Conjugate priors for exponential families. \newblock {\em Annals of Statistics}, 7(2):269--281. \bibitem[Faden, 1985]{Faden85} Faden, A. (1985). \newblock The existence of regular conditional probabilities: Necessary and sufficient conditions. \newblock {\em The Annals of Probability}, 13(1):288--298. \bibitem[Giry, 1982]{Giry82} Giry, M. (1982). \newblock A categorical approach to probability theory. \newblock In Banaschewski, B., editor, {\em Categorical Aspects of Topology and Analysis}, number 915 in Lect. Notes Math., pages 68--85. Springer, Berlin. \bibitem[Jacobs, 2013]{Jacobs13a} Jacobs, B. (2013). \newblock Measurable spaces and their effect logic. \newblock In {\em Logic in Computer Science}. IEEE, Computer Science Press. \bibitem[Jacobs, 2015]{Jacobs15d} Jacobs, B. (2015). \newblock New directions in categorical logic, for classical, probabilistic and quantum logic. \newblock {\em Logical Methods in Comp. Sci.}, 11(3). \newblock See \url{https://lmcs.episciences.org/1600}. \bibitem[Jacobs, 2017]{Jacobs17a} Jacobs, B. (2017). \newblock From probability monads to commutative effectuses. \newblock {\em Journ. of Logical and Algebraic Methods in Programming}, 94:200--237. \bibitem[Jacobs and Westerbaan, 2015]{JacobsW15a} Jacobs, B. and Westerbaan, A. (2015). \newblock An effect-theoretic account of {Lebesgue} integration. \newblock In Ghica, D., editor, {\em Math. Found. of Programming Semantics}, number 319 in Elect. Notes in Theor. Comp. Sci., pages 239--253. Elsevier, Amsterdam. \bibitem[Jacobs and Zanasi, 2016]{JacobsZ16} Jacobs, B. and Zanasi, F. (2016). \newblock A predicate/state transformer semantics for {Bayesian} learning. \newblock In Birkedal, L., editor, {\em Math. Found. of Programming Semantics}, number 325 in Elect. Notes in Theor. Comp. Sci., pages 185--200. Elsevier, Amsterdam. \bibitem[Koopman, 1936]{Koopman36} Koopman, B. (1936). \newblock On distributions admitting a sufficient statistic. \newblock {\em Trans. Amer. Math. Soc.}, 39:399--409. \bibitem[Panangaden, 2009]{Panangaden09} Panangaden, P. (2009). \newblock {\em Labelled {Markov} Processes}. \newblock Imperial College Press, London. \bibitem[Russell and Norvig, 2003]{RussellN03} Russell, S. and Norvig, P. (2003). \newblock {\em Artificial Intelligence. A Modern Approach}. \newblock Prentice Hall, Englewood Cliffs, NJ. \bibitem[Selinger, 2007]{Selinger07} Selinger, P. (2007). \newblock Dagger compact closed categories and completely positive maps (extended abstract). \newblock In Selinger, P., editor, {\em Proceedings of the 3rd International Workshop on Quantum Programming Languages (QPL 2005)}, number 170 in Elect. Notes in Theor. Comp. Sci., pages 139--163. Elsevier, Amsterdam. \newblock DOI \url{http://dx.doi.org/10.1016/j.entcs.2006.12.018}. \bibitem[Selinger, 2011]{Selinger11} Selinger, P. (2011). \newblock A survey of graphical languages for monoidal categories. \newblock In Coecke, B., editor, {\em New Structures in Physics}, number 813 in Lect. Notes Physics, pages 289--355. Springer, Berlin. \bibitem[Stoyanov, 2014]{Stoyanov14} Stoyanov, J. (2014). \newblock {\em Counterexamples in Probability}. \newblock Wiley, $2^{\textrm{nd}}$ rev. edition. \end{thebibliography} \end{document}