当前位置: 首页 > news >正文

佛山网站建设怎么做广州外贸网站建设公司价格

佛山网站建设怎么做,广州外贸网站建设公司价格,网站如何做mip,重庆市建筑工程【已解决】TD算法超详细解释和实现#xff08;Sarsa#xff0c;n-step Sarsa#xff0c;Q-learning#xff09;一篇文章看透彻#xff01; 郑重声明#xff1a;本系列内容来源 赵世钰(Shiyu Zhao)教授的强化学习数学原理系列#xff0c;本推文出于非商业目的分享个人学习…【已解决】TD算法超详细解释和实现Sarsan-step SarsaQ-learning一篇文章看透彻 郑重声明本系列内容来源 赵世钰(Shiyu Zhao)教授的强化学习数学原理系列本推文出于非商业目的分享个人学习笔记和心得。如有侵权将删除帖子。原文链接https://github.com/MathFoundationRL/Book-Mathmatical-Foundation-of-Reinforcement-Learning 上一节我们讲到Robbins-Monro Algorithm算法解决了下面的这个求期望的问题本节我们把问题稍微复杂化一点。 看下边这个期望的计算 假设我们可以获得随机变量R,XR,XR,X的样本那么可以定义下面的函数 其实也就是是把之前的一个随机变量变成了一个多元随机变量的函数。下面我们展示这个例子其实和TD算法的表达很相似了。 TD learning of state values 本节所指的TD算法特指用于估测状态价值的经典TD算法。 算法 这个状态价值的期望形式的表示有时候被称为贝尔曼期望等式其实是贝尔曼方程的另一种表达它是设计和分析TD算法的重要工具。那怎么去解这个方程呢这就用到了我们前面小节讲到的RM算法以及随机梯度下降的思想。这里需要注意每次更新是一次更新一个状态的价值。 推导 通过应用RM算法求解贝尔曼方程我们可以导出TD算法。 在了解了RM算法的前提下上面的算法并不难理解。另外RM算法关于状态价值的解7.5式有两个假设值得考虑 显然假设1)和2)在实践中都无法实现对于1是因为实践中的数据都是一个个的episodess在其中只是被抽样到的一个不能确保每一个episodes都是从s开始的对于2我们怎么可以在事先知道其他状态的价值呢要知道每一个状态的价值都是被算法估测出来的。 如何在RM算法中去除这两个假设的约束呢 也就是说RM算法中用的数据以s为起点把第k次抽样s′s^\primes′ 得到的 sk′s_k^\primesk′​ 作为下一个状态而这样并不能充分利用全局的状态来进行价值的估测要知道我们得到的数据是一个个episodes每次只用episodes的开头部分的数据是不可行的。 下面这个地方说明了k和t 的不同 所以我们修改了算法使用的序列样本数据的形式为{(st,rt1,st1)}\left\{\left(s_t, r_{t1}, s_{t1}\right)\right\}{(st​,rt1​,st1​)} 另一个关于状态价值的改动就不必说了因为算法更新过程中的状态价值并不是策略的状态价值函数一个是进行时态一个是完成时态。 性质 首先我们要深入分析一下TD算法的表达式 第二TD算法也有局限性它只能估测状态的价值为了找到最优的策略我们还需要进一步计算动作价值然后做策略改进。当然啦在系统模型不可知的时候我们可以直接去估测动作价值 第三TD算法和MC算法都是无模型的TD算法和MC相比的优势和劣势是什么呢我们可以从下表看到 PS个人觉得这个表意义重大值得反复观看 区分了online learning 和offline learning区分了持续性任务和周期性任务区分了自举和非自举分析了MC和TD算法的方差比较和成因赵世钰老师牛逼 收敛性分析 略了 TD learning of action values: Sarsa 算法 sarsa算法的作用是去估计动作的价值。 在第t步只有当前的状态-动作对被更新。 SARSA的名字来源 Sarsa is the abbreviation of state-action-reward-state-action. SARSA 和 TD的关系 We can obtain Sarsa by replacing the state value estimate v(s) in the TD algorithm with the action value estimate q(s,a). As a result, Sarsa is nothing but an action-value version of the TD algorithm. Sarsa is a stochastic approximation algorithm solving the following equation: SARSA算法是贝尔曼方程的状态-动作价值形式。 SARSA算法的收敛性对学习率 α\alphaα 有要求 实现 先捋一下思路RL算法的目的是找最优的策略。而SARSA算法目前只能求解最优的状态-动作价值。收基于模型的策略迭代算法的启发我们可以把策略迭代引入SARSA分成两步来找最优策略The first is to update the q-value of the visited state-action pair. The second is to update the policy as an ε-greedy one.有两点值得注意 第一点之前值迭代和一些基于模型的策略迭代在每一次迭代之后是更新了所有的状态的价值的而SARSA在每一步仅仅更新一个状态-动作对的价值然后立马状态sts_tst​的策略就被随之更新。尽管啊这样的更新策略可能是不够精确的。 第二点 policy 更新过程是 ε-greedy 的从而保证所有的状态-动作对都被访问到。 我们来看一个例子无图无真相 实验设定如下 下图显示的是SARSA最终算出来的策略注意this policy can successfully reach the target state from the starting state. However, the policies for other states may not be optimal. That is because the other states are not of interest and every episode ends when the target state is reached. If we need to find the optimal policies for all states, we should ensure all the states are visited sufficiently many times by, for example, starting episodes from different states. 说白了就是初始的那个状态唯我独尊但是对其他状态作为初始状态的情况这个生成的策略并不是最优的。要想对所有状态找最优策略那么我们需要把每一个状态作为开始状态得到充分多的状态动作对作为episods训练数据。 右上图可以看到每一个episode的sum reward在逐渐增高。从右下图可以看到episodes的长度在逐渐变短。 TD learning of action values: Expected Sarsa 原理 实现 在算法实现上面除了q-value的更新过程不一样之外其他的和SARSA差不多。 实验 TD learning of action values: n-step Sarsa 和Sarsa的区别主要是discounted return项的展开的长度the different decomposition structure 的不一样 实现 注意要点 However, we can wait until time t n to update the q-value of (St,At). To that end, (7.14) can be modified to 区别就是我们通过等待n步再计算q值从而让它更准确一点。 总结 Specifically, if n is selected to be a large number, its performance is close to MC learning and hence has a large variance but a small bias. If n is selected to be small, its performance is close to Sarsa and hence has a relatively large bias due to the initial guess and relatively low variance.Finally, n-step Sarsa is also for policy evaluation. It can be combined with the policy improvement step to search for optimal policies. TD learning of optimal action values: Q-learning 放大招 SARSA只能是估测每个状态动作价值必须要借助策略迭代方法来求最优状态动作价值而By contrast, Q-learning can directly estimate optimal action values. 算法 Q: Q-learning的更新过程需要的数据是什么呢 A反正不需要at1a_{t1}at1​ Q-learning其实是在使用随机估测方法在计算下面的方程 很简单可以证明这个方程其实就是贝尔曼最优方程 展开上式得到 实现 虽然Q-learning 是off-policy这是由算法本身决定的但是它可以Work online 或者 offline这是和实现有关系的。 The on-policy version is the same as Sarsa except for the TD target in the q-value update step. In this case, the behavior policy is the same as the target policy, which is an ε-greedy policy. In the off-policy version, the episode is generated by the behavior policy πb which can be any policy. If we would like to sample experiences for every state-action pair, πb should be selected to be exploratory.意思是把学习率设置得比较大 Here, the target policy πT is greedy rather than ε-greedy. That is because the target policy is not used to generate episodes and hence is not required to be exploratory. Moreover, the off-policy version of Q-learning presented here is offline. That is because all the experience samples are collected first and then processed to estimate an optimal target policy. Of course, it can be modified to become online. Once a sample is collected, the sample can be used to update the q-value and target policy immediately. Nevertheless, the updated target policy is not used to generate new samples. online版本的Q-learning没有使用更新后的策略去生成样本进行迭代否则的话就是on-policy了 例子图示 Figure 7.2: Examples to demonstrate off-policy learning by Q-learning. The optimal policy and optimal state values are shown in (a) and (b), respectively. The behavior policy and the generated single episode are shown in © and (d), respectively. The estimated policy and the estimation error evolution are shown in (e) and (f), respectively. The cases with different initial q-value are shown in (g) and (h), respectively. 由于最优策略有多个所以Q-learning学到的e和原始的最优策略设定b并不一样。 由于Q-learning的自举特性初始状态价值的设定和收敛的速度是息息相关的。g和h的区别就可以看出来。 Figure 7.3: Performance of Q-learning given different non-exploratory behavior policies. The figures in the left column show the behavior policies. The figures in the middle column show the generated episodes following the corresponding behavior policies. The episode in every example has 100,000 steps. The figures in the right column show the evolution of the root-mean-squared error of the estimated state values, which are calculated from the estimated action values.上面这些图中间的那列是基于右边的策略生成的轨迹。 可以看出啊当行为策略不是探索的时候在7.3中是固定的学习的效果变得很差。比如状态值的误差对 ϵ\epsilonϵ的缩小很敏感。 A unified viewpoint It is worth mentioning that there are some methods to convert an on-policy learning algorithm to off-policy. Importance sampling is a widely used method1. Moreover, there are various extensions of the TD algorithms introduced in this chapter. For example, the TD(λ) method provides a more general and unified framework for TD learning. The TD algorithm in Section 7.6 is simply TD(0), a special case of TD(λ). Details of TD(λ) can be found in 2. 注意Q-learning is off-policy 答疑环节 时间差分TD这个词在TD learning中是什么意思 Every TD algorithm has a TD error, which represents the deficiency between the new sample and the current estimates. This deficiency is between two different time steps and hence called temporal difference. TD learning 算法只是在估计动作的价值它怎么可以用来找寻最优策略呢 To obtain an optimal policy, the value estimation process should interact with the policy update process. That is, after a value is updated, the corresponding policy should be updated. Then, a new sample generated by the updated policy is used to update values again. This is the idea of generalized policy iteration. TD learning 算法的收敛要求学习率逐渐收敛到0为什么实践中学习率被设置为常量呢 The fundamental reason is the policy to be evaluated is nonstationary. A TD learning algorithm like Sarsa aims to estimate the action values of a given fixed policy. In other words, it aims to evaluate the given policy. If the policy is stationary, the learning rate can decrease to zero gradually to ensure convergence with probability 1 as stated in Theorem 7.1. 在策略评估场景下我们假定策略不变However, when we put Sarsa in the context of optimal policy searching, the value estimation process keeps interacting with the policy update process. Therefore, the policy that Sarsa aims to evaluate keeps changing. In this case, we need to use a constant learning rate because, every time we have a new policy to evaluate, the learning rate cannot be too small. The drawback of a constant learning rate is that the value estimate may still fluctuate eventually. However, as long as the constant learning rate is sufficiently small, the fluctuation would not jeopardize the convergence.在最优策略寻找场景下策略在不断地更新面对新的策略我们不能让学习率太小了。缺点呢就是状态价值的估计值会一直震荡。这个情况可以通过把学习率设置得比较小来实现。 Should we estimate the optimal policies for all states or a subset of states? It depends on the task. One may have noticed that some tasks in the illustrative examples in this chapter are not to find the optimal policies for all states. Instead, they only need to find a good path from a specific starting state to the target state. Such a task is not data demanding because we do not need to visit every state-action pair sufficiently many times. As shown in the examples in Figure 7.1, even if some states are not visited, we can still obtain a good path. It, however, must be noted that the path is only good but not guaranteed to be optimal. That is because not all state-action pairs have been explored and there may be better paths unexplored. Nevertheless, given sufficient data, there is a high probability for us to find a good or locally optimal path. By locally optimal, we mean the path is optimal within a neighborhood of the path.重点理解一下locally optimal这个概念 为啥 Q-learning是off-policy的而别的TD算法都是 on-policy的而根本的原因是什么 Q-learning aims to solve the Bellman optimality equation. The other TD algorithms are on-policy because they aim to solve the Bellman equation of a given policy.Since the Bellman optimality equation does not depend on any specific policy, it can use the experience samples generated by any other policies to estimate the optimal policy. However, the Bellman equation depends on a given policy, solving it of course requires the experience samples generated by the given policy. 看明白了背后的原因是Bellman optimality equation 和 Bellman equation 在数学形式上的差别 T. C. Hesterberg, Advances in importance sampling. PhD Thesis, Stanford Univer- sity, 1988. ↩︎ R. S. Sutton, “Learning to predict by the methods of temporal differences,” Machine Learning, vol. 3, no. 1, pp. 9–44, 1988. ↩︎
http://www.dnsts.com.cn/news/259795.html

相关文章:

  • 长沙官网网站制作公司wordpress 文章缩略图
  • 深圳网站建设便捷建设银行徐州分行网站
  • 北京外包做网站如何报价wordpress的注册文件在哪个文件夹
  • seo网站快速排名软件广州哪家做网站最好
  • 网站首页导航栏怎么做网站建设网站模版
  • flash网站 seo网站上线如何做公司名字
  • 云南建设学校网站阿里巴巴可以做网站吗
  • 网站开发首选畅扬科技top wang域名做网站好
  • 做网站之前需要准备什么条件易网网站
  • 上海市住房和城乡建设网站网站源代码怎么下载
  • 网站主题旁边的图标怎么做欧美平面设计网站
  • 建设银行无锡分行网站网站建设每月工作多少
  • 企业建站公司案例linode vps wordpress 两个
  • 高埗仿做网站绍兴市柯桥区建设局网站
  • 临沂网站案例海安环评在哪个网站做
  • 徐州云建站模板安徽六安天气预报
  • 长安网站建设公司做淘宝客网站教程
  • 哪个网站亲子游做的好小视频制作软件app
  • 专业网站建站费用建设网站的方法
  • 佛山做网站那家好建设银行信用卡申请官方网站
  • 网站建设基础教程视频网站专题报道怎么做
  • 泉州seo网站排名全国企业信息查询官网系统
  • 建站公司用的服务器做雷达干扰的网站
  • 网站建设规模与类别wordpress怎样修改页脚版权信息
  • 网站建设服务哪家有上海有什么seo公司
  • 加盟网站制作公司私人做网站收费
  • 嘉峪关做网站怎么自己创建网站
  • 黑龙江建设网网上办事大厅网站架构优化 amp
  • h5游戏网站入口网站版本功能列表
  • 厦门营销型网站东莞网站推广建设