논문 리뷰/Reinforcement Learning

    [논문 리뷰] Asynchronous Methods for Deep Reinforcement Learning (A3C)

    [논문 리뷰] Asynchronous Methods for Deep Reinforcement Learning (A3C)

    이번 논문에서는 강화학습을 비동기적이게 학습을 하게 만든 논문을 들고 왔다. 이 논문의 특이점이라고 한다면 보통의 학습에서 쓰이는 GPU를 사용하지 않고 CPU 코어들을 통한 병렬학습을 한다는 것이다. 이를 통해 Atari 벤치마크에서 새로운 기록을 세웠고 다른 도메인에서도 좋은 결과를 보여주는 모습이다. [1602.01783] Asynchronous Methods for Deep Reinforcement Learning (arxiv.org) Asynchronous Methods for Deep Reinforcement Learning We propose a conceptually simple and lightweight framework for deep reinforcement learning that..

    [논문 리뷰] Rainbow: Combining Improvements in Deep Reinforcement Learning (Rainbow DQN)

    [논문 리뷰] Rainbow: Combining Improvements in Deep Reinforcement Learning (Rainbow DQN)

    [1710.02298] Rainbow: Combining Improvements in Deep Reinforcement Learning (arxiv.org) Rainbow: Combining Improvements in Deep Reinforcement Learning The deep reinforcement learning community has made several independent improvements to the DQN algorithm. However, it is unclear which of these extensions are complementary and can be fruitfully combined. This paper examines six extensions to the ..

    [논문 리뷰] Noisy Networks for Exploration (NoisyNet)

    [논문 리뷰] Noisy Networks for Exploration (NoisyNet)

    [1706.10295] Noisy Networks for Exploration (arxiv.org) Noisy Networks for Exploration We introduce NoisyNet, a deep reinforcement learning agent with parametric noise added to its weights, and show that the induced stochasticity of the agent's policy can be used to aid efficient exploration. The parameters of the noise are learned with grad arxiv.org 이번 논문에는 DQN에 있는 fully connected layer에 param..

    [논문 리뷰] Prioritized Experience Replay (PER)

    [논문 리뷰] Prioritized Experience Replay (PER)

    [1511.05952] Prioritized Experience Replay (arxiv.org) Prioritized Experience Replay Experience replay lets online reinforcement learning agents remember and reuse experiences from the past. In prior work, experience transitions were uniformly sampled from a replay memory. However, this approach simply replays transitions at the same frequ arxiv.org 이 논문은 DQN의 uniformly sampled experience replay..

    [논문 리뷰] Dueling Network Architectures for Deep Reinforcement Learning (Dueling DQN)

    [논문 리뷰] Dueling Network Architectures for Deep Reinforcement Learning (Dueling DQN)

    [1511.06581] Dueling Network Architectures for Deep Reinforcement Learning (arxiv.org) Dueling Network Architectures for Deep Reinforcement Learning In recent years there have been many successes of using deep representations in reinforcement learning. Still, many of these applications use conventional architectures, such as convolutional networks, LSTMs, or auto-encoders. In this paper, we pres..