(づ•ᴥ•)づ┬─┬

Notes

A few notes on directed information

TL;DR: Notes on directed information, inspired from the use of directed information in Abel, David et al. “Plasticity as the Mirror of Empowerment.” ArXiv abs/2505.10361 (2025) – I want to eventually write some notes on this work as well.

Mutual Information

Quick recap of mutual information, for more see the wiki.

Suppose we have random variables X,Y with values over some space 𝒳×𝒴. Then, the mutual information (MI) captures, on average, how much information we learn about X if we observe Y. MI is defined as

I(X;Y)=x,yp(x,y)logp(x,y)p(x)p(y)=KL(p(x,y)p(x)p(y)).

Note the tension between the joint p(x,y) and the marginal product p(x)p(y).

There are lots of useful relationships around I, but here I’ll only mention a few:

  1. Symmetry: I(X;Y)=I(Y;X).Because of symmetry, there is no notion of causality or time in MI.
  2. Non-negativity: I(X;Y)0.
  3. Expressing in terms of entropy / relative entropy: I(X;Y)=H(X)H(X|Y)=H(Y)H(Y|X).

From 3., we can also get that if X and Y are independent, then I(X;Y)=0, i.e., there’s no information to be learnt from Y about X. On the other side, if, for example, X=Y, then H(X|Y)=0, so I(X;X)=H(X), i.e., we have learned all there is to know.

Directed Information

To introduce directed information, we will use notation from the original paper by Massey, 19901.

For n1, Xn:=[X1,,Xn], each component being a discrete random variable. Values of those will be denoted by xn:=[x1,,xn]. We use square brackets to denote that those objects are sequences of random variables.

In the original work1, Massey was studying the behavior of a discrete channel with input (X) and output (Y) random variables. One can make all sorts of assumptions on how an input variable Xk affects the corresponding output variable Yk. For example, we can assume that X1Y1, then X2Y2, (arrows here denote dependence) and so forth, each generation being independent of the previous one.

A bit more interesting is to assume that each output variable plays a role into the generation of the next one, the DAG looking something like this:

Untitled-2025-11-06-1143

How much information do we get about YN if we know XN? We could try to answer this question with I(XN;YN), but this is a symmetric quantity and doesn’t take into account the order the variables enter the picture.

This is where directed information (DI) can help:

I(XNYN):=i=1NI(Xi;Yi|Yi1).

Here I(Xi;Yi|Yi1) is regular conditional MI. DI captures how information flows through the channel, from input to output.

There are a few nice properties of DI as well, here are some I liked:

  1. I(XNYN)I(YNXN): this asymmetry is characteristic of DI. For example, one can show it by considering n=2 and Y2=X2 and Y1=X1.
  2. I(XNYN)I(XN;YN): each term of the DI sum contains the partial sequence Xi and replacing by the full sequence XN can only increase mutual information, which then gives the result (by means of a telescopic sum). Equality holds when there is no feedback from output back to input.
  3. Conservation of information! I(XN;YN)=I(XNYN)+I(YN1XN), note that the second term starts from YN1. Mutual information captures both the forward and the backward flow of information, hence the symmetry.

Proving conservation of information (sketch)

There are various ways to show 3 for DI. For example, for an appropriate definition of conditional entropy H(XNYN) (called causally conditioned entropy), we get an equivalent formula to MI’s 3:

I(XNYN)=H(YN)H(YNXN).

Then it is sufficient to show that

H(XN,YN)=H(XNYN)+H(YNXN),

which requires the decomposition rule for causal conditioning (denoted by )

P(xN,yN)=P(xNyN1)P(yNxN)

and the definition of joint entropy as

H(XN,YN):=xN,yNP(xN,yN)logP(xN,yN).

The I(YN1XN) term

The backward I(YN1XN) can be interpreted as the strength of feedback: how much past outputs influence future inputs.

Generalized Directed Information

To handle causal interactions that may begin or end at different times, we can generalize Massey’s directed information to arbitrary causal windows. Following Abel et al. (2025), the Generalized Directed Information (GDI) between sequences Xa:b and Yc:d is defined as

I(Xa:bYc:d):=i=max(a,c)dI(Xa:min(b,i);YiX1:a1,Y1:i1).

This quantity measures how much information flows from the segment Xa:b into the segment Yc:d, respecting the causal structure of the process. When a=c=1 and b=d=n, it reduces to Massey’s original directed information I(XnYn).

GDI also satisfies an entropy–difference identity, sometimes called the Kramer Decomposition:

I(Xa:bYc:d)=H(Ymax(a,c):dX1:a1)i=max(a,c)dH(YiY1:i1,Xa:min(b,i)).

This form generalizes the standard relation I(XnYn)=H(Yn)H(Yn|Xn), extending it to arbitrary causal intervals.

  1. Massey, J. (1990, November). Causality, feedback and directed information. In Proc. Int. Symp. Inf. Theory Applic.(ISITA-90) (Vol. 2, p. 1).