Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CPE Functionaility - Purpose of additional CPE Q-Networks? #456

Open
tfurmston opened this issue Apr 19, 2021 · 0 comments
Open

CPE Functionaility - Purpose of additional CPE Q-Networks? #456

tfurmston opened this issue Apr 19, 2021 · 0 comments

Comments

@tfurmston
Copy link

Hi,

I have been reading through the code and I have a few questions regarding the CPE functionality.

In particular, with the DQN model you have separate Q-networks, both normal and target, dedicated just for CPE that I would like to better understand. At present their purpose is not really clear to me. In particular, what is the purpose of these additional networks over and above the standard Q-networks of the DQN model?

In the function _calculate_cpes, which is part of the RLTrainer class. Reading this function it seems that update the networks q_network_cpe and q_network_cpe_target to model not only the reward, but any additional metrics that could be on interest in CPE.

Am I right in thinking that performing CPE on these additional metrics is the main reason for these additional networks? Put another way, if one were only interested in performing CPE on the reward itself, would using the standard Q-networks of the DQN model suffice?

Thanks

@tfurmston tfurmston changed the title CPE Functionaility CPE Functionaility - Purpose of additional CPE Q-Networks? Apr 19, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant