# Q-learning

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
34,196pages on
this wiki

Q-learning is a reinforcement learning technique that works by learning an action-value function that gives the expected utility of taking a given action in a given state and following a fixed policy thereafter. A strength with Q-learning is that it is able to compare the expected utility of the available actions without requiring a model of the environment. A recent variation called delayed-Q learning has shown substantial improvements, bringing PAC bounds to Markov decision processes.

## Algorithm Edit

The problem model consists of an agent, states S and a number of actions per state A. By performing an action a, the agent can move from state to state. Each state provides the agent a reward (a real or natural number) or punishment (a negative reward). The goal of the agent is to maximize its total reward. It does this by learning which action is optimal for each state.

The algorithm therefore has a function which calculates the Quality of a state-action combination:

$Q: S \times A \to \mathbb{R}$

Before learning has started, Q returns a fixed value, chosen by the designer. Then, each time the agent is given a reward (the state has changed) new values are calculated for each combination of a state s from S, and action a from A. The core of the algorithm is a simple value iteration update. It assumes the old value and makes a correction based on the new information.

$Q(s_t,a_t) \leftarrow \underbrace{Q(s_t,a_t)}_{old~value} + \underbrace{\alpha_t(s_t,a_t)}_{learning~rate} \times [\overbrace{\underbrace{r_{t+1}}_{reward} + \underbrace{\gamma}_{discount~factor} \underbrace{\max_{a}Q(s_{t+1}, a)}_{max~future~value}}^{expected~discounted~reward} - \overbrace{Q(s_t,a_t)}^{old~value}]$

Where $r_t$ is the reward given at time $t$, $\alpha_t(s, a)$ ($0 < \alpha \le 1$) the learning rates, may be the same value for all pairs. The discount factor $\gamma$ is such that $0 \le \gamma < 1$

The above formula is equivalent to:

$Q(s_t,a_t) \leftarrow Q(s_t,a_t)(1-\alpha_t(s_t,a_t)) + \alpha_t(s_t,a_t) [r_{t+1} + \gamma \max_{a}Q(s_{t+1}, a)]$

## Influence of variables on the algorithm Edit

### Learning rate Edit

The learning rate determines to what extent the newly acquired information will override the old information. A factor of 0 will make the agent not learn anything, while a factor of 1 would make the agent consider only the most recent information.

### Discount factor Edit

The discount factor determines the importance of future rewards. A factor of 0 will make the agent "opportunistic" by only considering current rewards, while a factor of 1 will make it strive for a long-term high reward.

## ImplementationEdit

Q-learning at its simplest uses tables to store data. This very quickly loses viability with increasing levels of complexity of the system it is monitoring/controlling. One answer to this problem is to use an (adapted) artificial neural network as a function approximator, as demonstrated by Tesauro in his Backgammon playing temporal difference learning research. An adaptation of the standard neural network is required because the required result (from which the error signal is generated) is itself generated at run-time.