Difference between revisions of "Allen's REINFORCE notes"

From Humanoid Robots Wiki
Jump to: navigation, search
(Overview)
Line 16: Line 16:
 
‎<syntaxhighlight lang="bash" line>
 
‎<syntaxhighlight lang="bash" line>
 
Initialize neural network with input dimensions = observation dimensions and output dimensions = action dimensions
 
Initialize neural network with input dimensions = observation dimensions and output dimensions = action dimensions
For \# of episodes:
+
For each episode:
 
   While not terminated:
 
   While not terminated:
 
     Get observation from environment
 
     Get observation from environment
Line 24: Line 24:
 
     Step environment using action and store reward
 
     Step environment using action and store reward
 
   Calculate loss over entire trajectory as function of probabilities and rewards
 
   Calculate loss over entire trajectory as function of probabilities and rewards
 +
  Recall loss functions are differentiable with respect to each parameter - thus, calculate how changes in parameters correlate with changes in the loss
 +
  Based on the loss, use a gradient descent policy to update weights
 +
 
</syntaxhighlight>
 
</syntaxhighlight>
  
 
=== Loss Function ===
 
=== Loss Function ===

Revision as of 00:15, 25 May 2024

Allen's REINFORCE notes

Links

Motivation

Recall that the objective of Reinforcement Learning is to find an optimal policy which we encode in a neural network with parameters . is a mapping from observations to actions. These optimal parameters are defined as . Let's unpack what this means. To phrase it in english, this is basically saying that the optimal policy is one such that the expected value of the total reward over following a trajectory () determined by the policy is the highest over all policies.

Overview

 1 Initialize neural network with input dimensions = observation dimensions and output dimensions = action dimensions
 2 For each episode:
 3   While not terminated:
 4     Get observation from environment
 5     Use policy network to map observation to action distribution
 6     Randomly sample one action from action distribution
 7     Compute logarithmic probability of that action occurring
 8     Step environment using action and store reward
 9   Calculate loss over entire trajectory as function of probabilities and rewards
10   Recall loss functions are differentiable with respect to each parameter - thus, calculate how changes in parameters correlate with changes in the loss
11   Based on the loss, use a gradient descent policy to update weights

Loss Function