-
Notifications
You must be signed in to change notification settings - Fork 230
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RLLIB QMIX example does not work #12
Comments
cc @ericl |
I found this problem mainly from enabling |
I was able to run the RLlib's QMIX in the Starcraft2 env. However, the policy does not converge. |
I forgot how to make it converge. I recommend you to use pymarl instead of rllib if you want to explore some research ideas. |
Hi, where to set self.double.q=False How to fix it , and run well? Thanks |
@xiaoToby One key difference between the default PyMARL implementation and rllib implementation of QMIX for SMAC is that PyMARL uses the true overall global state but rllib only uses the per-agent observation as the global state in the monotonic mixing network. So you'd need to modify the default implementation and extract the true global state from the environment so that the mixing network can use it. |
|
There appears to be a problem when using a masked action space with the QMIX algorithm. I think the qmix_policy_graph expects there to be at least one valid action at all times.
Full traceback is below:
The text was updated successfully, but these errors were encountered: