Misplaced Pages

Blackwell's informativeness theorem

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Information theorem "Blackwell order" redirects here. For other uses, see Rao-Blackwell Theorem.

In the mathematical subjects of information theory and decision theory, Blackwell's informativeness theorem is an important result related to the ranking of information structures, or experiments. It states that there is an equivalence between three possible rankings of information structures: one based in expected utility, one based in informativeness, and one based in feasibility. This ranking defines a partial order over information structures known as the Blackwell order, or Blackwell's criterion.

The theorem states equivalent conditions under which any expected utility maximizing decision maker prefers information structure σ {\displaystyle \sigma } over σ {\displaystyle \sigma '} , for any decision problem. The result was first proven by David Blackwell in 1951, and generalized in 1953.

Setting

Decision making under uncertainty

A decision maker faces a set of possible states of the world Ω {\displaystyle \Omega } and a set of possible actions A {\displaystyle A} to take. For every ω Ω {\displaystyle \omega \in \Omega } and a A {\displaystyle a\in A} , her utility is u ( ω , a ) {\displaystyle u(\omega ,a)} . She does not know the state of the world ω {\displaystyle \omega } , but has a prior probability p : Ω [ 0 , 1 ] {\displaystyle p:\Omega \rightarrow } for every possible state. For every action she takes, her expected utility is

ω Ω u ( a , ω ) p ( ω ) {\displaystyle \sum _{\omega \in \Omega }u(a,\omega )p(\omega )}

Given such prior p {\displaystyle p} , she chooses an action a A {\displaystyle a\in A} to maximize her expected utility. We denote such maximum attainable utility (the expected value of taking the optimal action) by

V ( p ) = max a A ω Ω u ( a , ω ) p ( ω ) {\displaystyle V(p)={\underset {a\in A}{\operatorname {max} }}\sum _{\omega \in \Omega }u(a,\omega )p(\omega )}

We refer to the data ( Ω , A , u , p ) {\displaystyle (\Omega ,A,u,p)} as a decision making problem.

Information structures

An information structure (or an experiment) can be seen as way to improve on the utility given by the prior, in the sense of providing more information to the decision maker. Formally, an information structure is a tuple ( S , σ ) {\displaystyle (S,\sigma )} , where S {\displaystyle S} is a signal space and σ : Ω Δ S {\displaystyle \sigma :\Omega \rightarrow \Delta S} is a function which gives the conditional probability σ ( s | ω ) {\displaystyle \sigma (s|\omega )} of observing signal s S {\displaystyle s\in S} when the state of the world is ω {\displaystyle \omega } . An information structure can also be thought of as the setting of an experiment.

By observing the signal s {\displaystyle s} , the decision maker can update her beliefs about the state of the world ω {\displaystyle \omega } via Bayes' rule, giving the posterior probability

π ( ω | s ) = p ( ω ) σ ( s | ω ) π ( s ) {\displaystyle \pi (\omega |s)={\frac {p(\omega )\sigma (s|\omega )}{\pi (s)}}}

where π ( s ) := ω Ω p ( ω ) σ ( s | ω ) {\displaystyle \pi (s):=\sum _{\omega '\in \Omega }p(\omega ')\sigma (s|\omega ')} . By observing the signal s {\displaystyle s} and updating her beliefs with the information structure ( S , σ ) {\displaystyle (S,\sigma )} , the decision maker's new expected utility value from taking the optimal action is

V ( π , s ) = max a A ω Ω u ( a , ω ) π ( ω | s ) {\displaystyle V(\pi ,s)={\underset {a\in A}{\operatorname {max} }}\sum _{\omega \in \Omega }u(a,\omega )\pi (\omega |s)}

and the "expected value of ( S , σ ) {\displaystyle (S,\sigma )} " for the decision maker (i.e., the expected value of taking the optimal action under the information structure) is defined as

W ( σ ) = s S V ( π , s ) π ( s ) {\displaystyle W(\sigma )=\sum _{s\in S}V(\pi ,s)\pi (s)}

Garbling

If two information structures ( S , σ ) {\displaystyle (S,\sigma )} and ( S , σ ) {\displaystyle (S,\sigma ')} have the same underlying signal space, we abuse some notation and refer to σ {\displaystyle \sigma } and σ {\displaystyle \sigma '} as information structures themselves. We say that σ {\displaystyle \sigma '} is a garbling of σ {\displaystyle \sigma } if there exists a stochastic map (for finite signal spaces S {\displaystyle S} , a Markov matrix) Γ : S S {\displaystyle \Gamma :S\rightarrow S} such that

σ = Γ σ {\displaystyle \sigma '=\Gamma \sigma }

Intuitively, garbling is a way of adding "noise" to an information structure, such that the garbled information structure is considered to be less informative.

Feasibility

A mixed strategy in the context of a decision-making problem is a function α : S Δ A {\displaystyle \alpha :S\rightarrow \Delta A} which gives, for every signal s S {\displaystyle s\in S} , a probability distribution α ( a | s ) {\displaystyle \alpha (a|s)} over possible actions in A {\displaystyle A} . With the information structure ( S , σ ) {\displaystyle (S,\sigma )} , a strategy α {\displaystyle \alpha } induces a distribution over actions α σ ( a | ω ) {\displaystyle \alpha _{\sigma }(a|\omega )} conditional on the state of the world ω {\displaystyle \omega } , given by the mapping

ω α σ ( a | ω ) = s S α ( a | s ) σ ( s | ω ) Δ A {\displaystyle \omega \mapsto \alpha _{\sigma }(a|\omega )=\sum _{s\in S}\alpha (a|s)\sigma (s|\omega )\in \Delta A}

That is, α σ ( a | ω ) {\displaystyle \alpha _{\sigma }(a|\omega )} gives the probability of taking action a A {\displaystyle a\in A} given that the state of the world is ω Ω {\displaystyle \omega \in \Omega } under information structure ( S , σ ) {\displaystyle (S,\sigma )} – notice that this is nothing but a convex combination of the α ( a | s ) {\displaystyle \alpha (a|s)} with weights σ ( s | ω ) {\displaystyle \sigma (s|\omega )} . We say that α σ ( a | ω ) {\displaystyle \alpha _{\sigma }(a|\omega )} is a feasible strategy (or conditional probability over actions) under ( S , σ ) {\displaystyle (S,\sigma )} .

Given an information structure ( S , σ ) {\displaystyle (S,\sigma )} , let

Φ σ = { α σ ( a | ω ) {\displaystyle \Phi _{\sigma }=\{\alpha _{\sigma }(a|\omega )} α : S Δ A } {\displaystyle \alpha :S\rightarrow \Delta A\}}

be the set of all conditional probabilities over actions (i.e., strategies) that are feasible under ( S , σ ) {\displaystyle (S,\sigma )} .

Given two information structures ( S , σ ) {\displaystyle (S,\sigma )} and ( S , σ ) {\displaystyle (S,\sigma ')} , we say that σ {\displaystyle \sigma } yields a larger set of feasible strategies than σ {\displaystyle \sigma '} if

Φ σ Φ σ {\displaystyle \Phi _{\sigma '}\subset \Phi _{\sigma }}

Statement

Blackwell's theorem states that, given any decision-making problem ( Ω , A , u , p ) {\displaystyle (\Omega ,A,u,p)} and two information structures σ {\displaystyle \sigma } and σ {\displaystyle \sigma '} , the following are equivalent:

  1. W ( σ ) W ( σ ) {\displaystyle W(\sigma ')\leq W(\sigma )} : that is, the decision maker attains a higher expected utility under σ {\displaystyle \sigma } than under σ {\displaystyle \sigma '} .
  2. There exists a stochastic map Γ {\displaystyle \Gamma } such that σ = Γ σ {\displaystyle \sigma '=\Gamma \sigma } : that is, σ {\displaystyle \sigma '} is a garbling of σ {\displaystyle \sigma } .
  3. Φ σ Φ σ {\displaystyle \Phi _{\sigma '}\subset \Phi _{\sigma }} :, that is σ {\displaystyle \sigma } yields a larger set of feasible strategies than σ {\displaystyle \sigma '} .

Blackwell order

Definition

Blackwell's theorem allows us to construct a partial order over information structures. We say that σ {\displaystyle \sigma } is more informative in the sense of Blackwell (or simply Blackwell more informative) than σ {\displaystyle \sigma '} if any (and therefore all) of the conditions of Blackwell's theorem holds, and write σ B σ {\displaystyle \sigma '\preceq _{B}\sigma } .

The order B {\displaystyle \preceq _{B}} is not a complete one, and most experiments cannot be ranked by it. More specifically, it is a chain of the partially-ordered set of information structures.

Applications

The Blackwell order has many applications in decision theory and economics, in particular in contract theory. For example, if two information structures in a principal-agent model can be ranked in the Blackwell sense, then the more informative one is more efficient in the sense of inducing a smaller cost for second-best implementation.

References

  1. ^ de Oliveira, Henrique (2018). "Blackwell's informativeness theorem using diagrams". Games and Economic Behavior. 109: 126–131. doi:10.1016/j.geb.2017.12.008.
  2. ^ Kosenko, Andre (2021). "Algebraic Properties of Blackwell's Order and A Cardinal Measure of Informativeness". arXiv:2110.11399 .
  3. Blackwell, David (1951). "Comparison of Experiments". Second Berkeley Symposium on Mathematical Statistics and Probability: 2.
  4. Blackwell, David (1953). "Equivalent comparison of experiments". The Annals of Mathematical Statistics. 24 (2): 265–272. doi:10.1214/aoms/1177729032.
  5. Karni, Edi; Safra, Zvi (2022). "Hybrid decision model and the ranking of experiments". Journal of Mathematical Economics. 101. doi:10.1016/j.jmateco.2022.102700. S2CID 237370357.
  6. Grossman, Sanford J.; Hart, Oliver D. (1983). "An Analysis of the Principal-Agent Problem". Econometrica. 51 (1): 7–45. doi:10.2307/1912246. JSTOR 1912246.
  7. Laffont, Jean-Jacques; Martimort, David (2002). The Theory of Incentives: The Principal-Agent Model. Princeton University Press. ISBN 978-0691091846. JSTOR j.ctv7h0rwr.
Categories: