Reward Signal

Definition

Reward Signal ( )

The Reward Signal is a scalar value that the environment sends to the agent at each time step. It defines the goal of the Reinforcement Learning problem: the agent’s objective is to maximize the total cumulative reward (return) it receives over the long run.

The Reward Hypothesis

The Reward Hypothesis

“That all of what we mean by goals and purposes can be well thought of as the maximization of the expected value of the cumulative sum of a received scalar signal (called reward).” — Sutton & Barto

Mathematical Formulation

The agent maximizes the return :

Total Discounted Return

where:

  • is the discount factor, determining the present value of future rewards.

Key Properties

  • Immediate vs. Delayed: A reward is immediate feedback, but the agent must often sacrifice immediate reward to achieve higher long-term return.
  • Scalar: It must be a single number (though multi-objective RL exists, it usually boils down to a weighted scalar).
  • Environment Bound: The reward is defined by the environment, not the agent. The agent cannot “change the rules” to get more reward.

Intuition

What vs. How

The reward signal should tell the agent what you want it to achieve, not how to achieve it.

  • Bad Reward: Giving a chess AI points for taking pieces (it might take pieces but lose the game).
  • Good Reward: +1 for winning, -1 for losing, 0 otherwise.

Connections

Appears In