论文标题
基于增强学习的应用程序在云中自动化:调查
Reinforcement Learning-based Application Autoscaling in the Cloud: A Survey
论文作者
论文摘要
强化学习(RL)证明了在复杂不确定环境中自动解决决策问题的巨大潜力。 RL提出了一种计算方法,该方法允许在具有随机行为的环境中通过相互作用进行学习,在这种环境中,代理采取行动以最大程度地提高一些累积的短期和长期奖励。在游戏理论中显示了一些最令人印象深刻的结果,其中代理在GO或Starcraft 2等游戏中表现出超人的性能,这导致其在包括云计算在内的许多其他领域中逐渐采用。因此,RL似乎是在云中自动化的有前途的方法,因为可以学习透明(没有人类干预),动态(无静态计划)以及可自适应(不断更新)的资源管理政策来执行应用程序。与其他广泛使用的自动化策略相比,要考虑的三个重要方面是以临时方式定义的或基于元毛术的解决方案中定义的,这些策略是定义的。自动化利用云弹性,根据给定的优化标准优化应用程序的执行,该标准要求决定何时以及如何扩展/降低计算资源,以及如何将它们分配到即将到来的处理工作负载。考虑到云是一个动态和不确定的环境,必须采取此类行动。在此激励的情况下,许多作品将RL应用于云中的自动化问题。在这项工作中,我们对主要场所的这些建议进行了详尽的调查,并根据一组建议的分类法统一比较它们。我们还讨论了该地区的开放问题和前瞻性研究。
Reinforcement Learning (RL) has demonstrated a great potential for automatically solving decision-making problems in complex uncertain environments. RL proposes a computational approach that allows learning through interaction in an environment with stochastic behavior, where agents take actions to maximize some cumulative short-term and long-term rewards. Some of the most impressive results have been shown in Game Theory where agents exhibited superhuman performance in games like Go or Starcraft 2, which led to its gradual adoption in many other domains, including Cloud Computing. Therefore, RL appears as a promising approach for Autoscaling in Cloud since it is possible to learn transparent (with no human intervention), dynamic (no static plans), and adaptable (constantly updated) resource management policies to execute applications. These are three important distinctive aspects to consider in comparison with other widely used autoscaling policies that are defined in an ad-hoc way or statically computed as in solutions based on meta-heuristics. Autoscaling exploits the Cloud elasticity to optimize the execution of applications according to given optimization criteria, which demands to decide when and how to scale-up/down computational resources, and how to assign them to the upcoming processing workload. Such actions have to be taken considering that the Cloud is a dynamic and uncertain environment. Motivated by this, many works apply RL to the autoscaling problem in the Cloud. In this work, we survey exhaustively those proposals from major venues, and uniformly compare them based on a set of proposed taxonomies. We also discuss open problems and prospective research in the area.