Skip to content

[Question] Is DQN actually Double DQN?  #1737

@oliverc1623

Description

@oliverc1623

❓ Question

Hi, first off great work on making Stable-Baselines3 an excellent resource for deep reinforcement learning practitioners.

I noticed that your DQN implementation features a target q network which resembles Google's Deep Mind paper, Deep Reinforcement Learning with Double Q-learning. Meanwhile, Neural Fitted Q Iteration, by Riedmiller, calculates the target using the "current estimate" of the Q function. I am looking for clarification if DQN is truly a Double DQN. I hope to use this information to accurately hopefully implement prioritized experience replay based off of your DQN implementation.

Thanks,
Oliver

Checklist

Metadata

Metadata

Assignees

No one assigned

    Labels

    RTFMAnswer is the documentationduplicateThis issue or pull request already existsquestionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions