Allen's REINFORCE notes

From Humanoid Robots Wiki
Revision as of 00:15, 25 May 2024 by Allen12 (talk | contribs)
Jump to: navigation, search

Allen's REINFORCE notes

Links

Motivation

Recall that the objective of Reinforcement Learning is to find an optimal policy which we encode in a neural network with parameters . is a mapping from observations to actions. These optimal parameters are defined as . Let's unpack what this means. To phrase it in english, this is basically saying that the optimal policy is one such that the expected value of the total reward over following a trajectory () determined by the policy is the highest over all policies.

Overview

 1 Initialize neural network with input dimensions = observation dimensions and output dimensions = action dimensions
 2 For each episode:
 3   While not terminated:
 4     Get observation from environment
 5     Use policy network to map observation to action distribution
 6     Randomly sample one action from action distribution
 7     Compute logarithmic probability of that action occurring
 8     Step environment using action and store reward
 9   Calculate loss over entire trajectory as function of probabilities and rewards
10   Recall loss functions are differentiable with respect to each parameter - thus, calculate how changes in parameters correlate with changes in the loss
11   Based on the loss, use a gradient descent policy to update weights

Loss Function