%V 11 %A Budai ĂdĂĄm %A Csorba KristĂłf %D 2018 %P 17-20 %O Bibliogr.: 20. p. ; ĂśsszefoglalĂĄs angol nyelven %K SzĂĄmĂtĂĄstechnika - elĹadĂĄskivonat, Algoritmus - elĹadĂĄskivonat %L acta61754 %X One of the major challenges of artificial intelligence is to learn solving tasks which are considered to be challenging for even a human. Reinforcement learning is the most general learning framework among the three main type of learning methods: supervised, unsupervised and reinforcement learning. Most of the problems can easily fit into this framework. Experience shows that a lot of machine learning methods with non-linear function approximators suffers from instability regarding convergence. Reinforcement learning is more prone to diverge due to its ability to change the structure of its training data by modifying the way how it interacts with the environment. In this paper we investigate the divergence issue of DQN on the CartPole problem in terms of the algorithmâs parameters. Instead of the usual approach we do not focus on the successful trainings but instead we focus on the dark side where the algorithm fails on such an easy problem like CartPole. The motivation is to gain some further insight into the nature of the divergence issues on a specific problem. %T Deep reinforcement learning : a study of the CartPole problem %J Conference of PhD Students in Computer Science