REINFORCE algorithm with discounted rewards – where does gamma^t in the update come from?Reinforcement learning: understanding this derivation of n-step Tree Backup algorithmWhy do we normalize the discounted rewards when doing policy gradient reinforcement learning?How can we use the current rewards as a system input in the RUN time when working with Deep Q learning?Does self driving technology gain more from the data or the state-of-the-art algorithm?RL Policy Gradient: How to deal with rewards that are strictly positive?MDP - RL, Multiple rewards for the same state possible?Reinforcement learning: Discounting rewards in the REINFORCE algorithmHow does Q-Learning deal with mixed strategies?Pytorch: How to create an update rule the doesn't come from derivatives?

Do I have to worry about players making “bad” choices on level up?

Stark VS Thanos

Are Boeing 737-800’s grounded?

What's the metal clinking sound at the end of credits in Avengers: Endgame?

Feels like I am getting dragged in office politics

Why does processed meat contain preservatives, while canned fish needs not?

Has any spacecraft ever had the ability to directly communicate with civilian air traffic control?

"ne paelici suspectaretur" (Tacitus)

Historically, were women trained for obligatory wars? Or did they serve some other military function?

What gives an electron its charge?

Why does nature favour the Laplacian?

Build a trail cart

Can fracking help reduce CO2?

Any examples of headwear for races with animal ears?

You look catfish vs You look like a catfish

Find the coordinate of two line segments that are perpendicular

How to determine the actual or "true" resolution of a digital photograph?

Single Colour Mastermind Problem

Where did the extra Pym particles come from in Endgame?

How can I get precisely a certain cubic cm by changing the following factors?

How does a Swashbuckler rogue "fight with two weapons while safely darting away"?

When and why did journal article titles become descriptive, rather than creatively allusive?

Sci-fi novel series with instant travel between planets through gates. A river runs through the gates

What's the polite way to say "I need to urinate"?



REINFORCE algorithm with discounted rewards – where does gamma^t in the update come from?


Reinforcement learning: understanding this derivation of n-step Tree Backup algorithmWhy do we normalize the discounted rewards when doing policy gradient reinforcement learning?How can we use the current rewards as a system input in the RUN time when working with Deep Q learning?Does self driving technology gain more from the data or the state-of-the-art algorithm?RL Policy Gradient: How to deal with rewards that are strictly positive?MDP - RL, Multiple rewards for the same state possible?Reinforcement learning: Discounting rewards in the REINFORCE algorithmHow does Q-Learning deal with mixed strategies?Pytorch: How to create an update rule the doesn't come from derivatives?













1












$begingroup$


I'm looking at Sutton & Barto's rendition of the REINFORCE algorithm (from their book here, pg. 328).



Sutton & Barto's REINFORCE



I can't quite understand why there is $gamma^t$ on the last line. They say:




[..] in the boxed algorithms we are giving the algorithms for the general discounted [return] case. All of the ideas go through in the discounted case with appropriate adjustments [..] but involve additional complexity that distracts from the main ideas.




This doesn't quite make it clear for me.



I didn't find any other description of REINFORCE that would include this. Does anyone have an idea how the $gamma^t$ gets there?










share|improve this question









$endgroup$











  • $begingroup$
    Later rewards should be less important than current.
    $endgroup$
    – Carl Rynegardh
    Apr 8 at 12:58










  • $begingroup$
    Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
    $endgroup$
    – Tuetschek
    Apr 9 at 10:32















1












$begingroup$


I'm looking at Sutton & Barto's rendition of the REINFORCE algorithm (from their book here, pg. 328).



Sutton & Barto's REINFORCE



I can't quite understand why there is $gamma^t$ on the last line. They say:




[..] in the boxed algorithms we are giving the algorithms for the general discounted [return] case. All of the ideas go through in the discounted case with appropriate adjustments [..] but involve additional complexity that distracts from the main ideas.




This doesn't quite make it clear for me.



I didn't find any other description of REINFORCE that would include this. Does anyone have an idea how the $gamma^t$ gets there?










share|improve this question









$endgroup$











  • $begingroup$
    Later rewards should be less important than current.
    $endgroup$
    – Carl Rynegardh
    Apr 8 at 12:58










  • $begingroup$
    Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
    $endgroup$
    – Tuetschek
    Apr 9 at 10:32













1












1








1





$begingroup$


I'm looking at Sutton & Barto's rendition of the REINFORCE algorithm (from their book here, pg. 328).



Sutton & Barto's REINFORCE



I can't quite understand why there is $gamma^t$ on the last line. They say:




[..] in the boxed algorithms we are giving the algorithms for the general discounted [return] case. All of the ideas go through in the discounted case with appropriate adjustments [..] but involve additional complexity that distracts from the main ideas.




This doesn't quite make it clear for me.



I didn't find any other description of REINFORCE that would include this. Does anyone have an idea how the $gamma^t$ gets there?










share|improve this question









$endgroup$




I'm looking at Sutton & Barto's rendition of the REINFORCE algorithm (from their book here, pg. 328).



Sutton & Barto's REINFORCE



I can't quite understand why there is $gamma^t$ on the last line. They say:




[..] in the boxed algorithms we are giving the algorithms for the general discounted [return] case. All of the ideas go through in the discounted case with appropriate adjustments [..] but involve additional complexity that distracts from the main ideas.




This doesn't quite make it clear for me.



I didn't find any other description of REINFORCE that would include this. Does anyone have an idea how the $gamma^t$ gets there?







reinforcement-learning policy-gradients






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Apr 8 at 10:20









TuetschekTuetschek

61




61











  • $begingroup$
    Later rewards should be less important than current.
    $endgroup$
    – Carl Rynegardh
    Apr 8 at 12:58










  • $begingroup$
    Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
    $endgroup$
    – Tuetschek
    Apr 9 at 10:32
















  • $begingroup$
    Later rewards should be less important than current.
    $endgroup$
    – Carl Rynegardh
    Apr 8 at 12:58










  • $begingroup$
    Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
    $endgroup$
    – Tuetschek
    Apr 9 at 10:32















$begingroup$
Later rewards should be less important than current.
$endgroup$
– Carl Rynegardh
Apr 8 at 12:58




$begingroup$
Later rewards should be less important than current.
$endgroup$
– Carl Rynegardh
Apr 8 at 12:58












$begingroup$
Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
$endgroup$
– Tuetschek
Apr 9 at 10:32




$begingroup$
Thanks, I think you're right – or as a friend of mine put it, samples that have less following time steps should be less important. So the intuition is there, but I'll leave this here in case someone is able to do the math :-).
$endgroup$
– Tuetschek
Apr 9 at 10:32










0






active

oldest

votes












Your Answer








StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "557"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













draft saved

draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48872%2freinforce-algorithm-with-discounted-rewards-where-does-gammat-in-the-update-c%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown

























0






active

oldest

votes








0






active

oldest

votes









active

oldest

votes






active

oldest

votes















draft saved

draft discarded
















































Thanks for contributing an answer to Data Science Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid


  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.

Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48872%2freinforce-algorithm-with-discounted-rewards-where-does-gammat-in-the-update-c%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Marja Vauras Lähteet | Aiheesta muualla | NavigointivalikkoMarja Vauras Turun yliopiston tutkimusportaalissaInfobox OKSuomalaisen Tiedeakatemian varsinaiset jäsenetKasvatustieteiden tiedekunnan dekaanit ja muu johtoMarja VaurasKoulutusvienti on kestävyys- ja ketteryyslaji (2.5.2017)laajentamallaWorldCat Identities0000 0001 0855 9405n86069603utb201588738523620927

Which is better: GPT or RelGAN for text generation?2019 Community Moderator ElectionWhat is the difference between TextGAN and LM for text generation?GANs (generative adversarial networks) possible for text as well?Generator loss not decreasing- text to image synthesisChoosing a right algorithm for template-based text generationHow should I format input and output for text generation with LSTMsGumbel Softmax vs Vanilla Softmax for GAN trainingWhich neural network to choose for classification from text/speech?NLP text autoencoder that generates text in poetic meterWhat is the interpretation of the expectation notation in the GAN formulation?What is the difference between TextGAN and LM for text generation?How to prepare the data for text generation task

Is this part of the description of the Archfey warlock's Misty Escape feature redundant?When is entropic ward considered “used”?How does the reaction timing work for Wrath of the Storm? Can it potentially prevent the damage from the triggering attack?Does the Dark Arts Archlich warlock patrons's Arcane Invisibility activate every time you cast a level 1+ spell?When attacking while invisible, when exactly does invisibility break?Can I cast Hellish Rebuke on my turn?Do I have to “pre-cast” a reaction spell in order for it to be triggered?What happens if a Player Misty Escapes into an Invisible CreatureCan a reaction interrupt multiattack?Does the Fiend-patron warlock's Hurl Through Hell feature dispel effects that require the target to be on the same plane as the caster?What are you allowed to do while using the Warlock's Eldritch Master feature?