1) Markov decision
Markov决策
1.
The application of Markov decision in PMS has been studied a lot in foreign countries, but in our country it is contrary.
关于Markov决策在路面管理系统的应用方面,国外已进行了较多的研究,但我国在这方面的研究则较少。
2) Markov decision chain
Markov决策链
3) Markov decision processes
Markov决策过程
1.
In the optimal design and cootrol of preparative chromatographic processes,the obstacles appear when one tries to link the Wilson's framework of chromatographic theories based on partial differential equations(PDEs) with the Eulerian presentation to optimal control approaches based on discrete time states,such as Markov decision processes(MDP) or Model predictive control(MPC).
在制备色谱的优化设计和控制过程中,若试图把基于偏微分方程(PDE)-Eulerian描述的Wilson色谱理论框架和基于离散时间状态的优化控制方法(如Markov决策过程(MDP)和模型预测控制(MPC)等)衔接在一起时,就会出现明显的障碍。
2.
Based on continuous-time Markov decision processes, a power management policy optimization approach is proposed.
提出一种基于连续时间Markov决策过程的动态电源管理策略优化方法。
3.
An algorithm to optimize policy in call admission control is presented by using the method of Markov decision processes combined with performance potentials.
应用 Markov决策过程与性能势相结合的方法 ,给出了呼叫接入控制的策略优化算法。
4) Markov decision model
Markov决策模型
5) Markov decision process
Markov决策过程
1.
An algorithm of reinforcement learning for finite-horizon Markov decision processes;
一种有限时段Markov决策过程的强化学习算法
2.
Based on the theory of Markov decision processes,the policy iteration and online policy iteration algorithms for performance optimization are provided.
在Markov决策过程理论的基础上,给出了关于性能指标的策略迭代和在线策略迭代算法,并通过实例仿真说明该方法的优越性。
3.
Markov Decision Process(MDP)model is the general frame for solving reinforcement learning problems.
Markov决策过程(MDP)模型是解决激励学习问题的通用方法,而动态规划方法是Agent在具有Markov环境下与策略相关的值函数学习算法。
补充资料:Bayes决策函数
Bayes决策函数
1 1I
Ba酬湘决策函数【Baye戒an dedsi.加。比佣;E‘触以.。‘.pe山a别”l月a,勿拟”朋」 一个规则(函数)占枷),它对每个统计试验结果x给出一个决策占(x),取值于一给定的决策集内,它使期望损失达到最小,正如在统计问题的加卿方法(B ayesinnapProach)框架中所定义的那样.
说明:补充资料仅用于学习参考,请勿用于其它任何用途。
参考词条