Categories
  Encyclosphere.org ENCYCLOREADER
  supported by EncyclosphereKSF

Randomized algorithms as zero-sum games

From HandWiki - Reading time: 2 min

Randomized algorithms are algorithms that employ a degree of randomness as part of their logic. These algorithms can be used to give good average-case results (complexity-wise) to problems which are hard to solve deterministically, or display poor worst-case complexity. An algorithmic game theoretic approach can help explain why in the average case randomized algorithms may work better than deterministic algorithms.

Formalizing the game

Consider a zero-sum game between player A, whose strategies are deterministic algorithms, and player B, who's strategies are inputs for A's algorithms. The cost of a strategy profile is the running time of A's chosen algorithm on B's chosen input. Therefore, player A tries to minimize the cost, and player B tries to maximize it. In the world of pure strategies, for every algorithm that A chooses, B may choose the most costly input – this is the worst-case scenario, and can be found using standard complexity analysis.

But in the real world, inputs are normally not selected by an ‘evil opponent’ – rather, they come from some distribution over inputs. Since this is the case, if we allow the algorithms to also be drawn from some distribution, we may look at the game as one that allows mixed strategies. That is, each player chooses a distribution over its strategies.

Analysis

Incorporating mixed strategies into the game allows us to use von Neumann's minimax theorem:

[math]\displaystyle{ \min_R \max_D T(A,D) = \max_D \min_A T(A,D) \, }[/math]

where R is a distribution over the algorithms, D is a distribution over inputs, A is a single deterministic algorithm, and T(AD) is the average running time of algorithm a on input D. More specifically:

[math]\displaystyle{ T(A,D) = \,\underset{x \sim D}{\operatorname{E}}[T(A,X)]. \, }[/math]

If we limit the set of algorithms to a specific family (for instance, all deterministic choices for pivots in the quick sort algorithm), choosing an algorithm A from R is equivalent to running a randomized algorithm (for instance, running quick sort and randomly choosing the pivots at each step).

This gives us an insight on Yao's principle, which states that the expected cost of any randomized algorithm for solving a given problem, on the worst-case input for that algorithm, can be no better than the expected cost, for a worst-case random probability distribution on the inputs, of the deterministic algorithm that performs best against that distribution.




Licensed under CC BY-SA 3.0 | Source: https://handwiki.org/wiki/Software:Randomized_algorithms_as_zero-sum_games
1 | Status: cached on July 25 2024 10:26:18
↧ Download this article as ZWI file
Encyclosphere.org EncycloReader is supported by the EncyclosphereKSF