Reward Signal
Definition
Reward Signal ( )
The Reward Signal is a scalar value that the environment sends to the agent at each time step. It defines the goal of the Reinforcement Learning problem: the agent’s objective is to maximize the total cumulative reward (return) it receives over the long run.
The Reward Hypothesis
The Reward Hypothesis
“That all of what we mean by goals and purposes can be well thought of as the maximization of the expected value of the cumulative sum of a received scalar signal (called reward).” — Sutton & Barto
Mathematical Formulation
The agent maximizes the return :
Total Discounted Return
where:
- is the discount factor, determining the present value of future rewards.
Key Properties
- Immediate vs. Delayed: A reward is immediate feedback, but the agent must often sacrifice immediate reward to achieve higher long-term return.
- Scalar: It must be a single number (though multi-objective RL exists, it usually boils down to a weighted scalar).
- Environment Bound: The reward is defined by the environment, not the agent. The agent cannot “change the rules” to get more reward.
Intuition
What vs. How
The reward signal should tell the agent what you want it to achieve, not how to achieve it.
- Bad Reward: Giving a chess AI points for taking pieces (it might take pieces but lose the game).
- Good Reward: +1 for winning, -1 for losing, 0 otherwise.
Connections
- Input for: Temporal Difference Learning and Q-Learning
- Defines: Optimal Policy
- Linked to: Model of the Environment ()