-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
U_net_graph_dict key, Ising model #11
Comments
can you give me some more details of what you have tried? |
i use after generation of IsingModel data, i use the command
and there are two errors:
seems need "U_net_graph_dict", but "U_net_graph_dict = None" set by SolutionDataset_InMemory.getitem() function |
Do you have the latest version of the code? The default for --T_target is 0 because this is the target temperature for CO problems. But for the Ising model, a target temperature of 0 is not supported. Usually, one considers the Ising model at a target temperature T >> 0, so you should set --T_target to a higher temperature. |
The U_Net graph mode is a graph unet that I have tried out and it is depreciated as it did not work well. The Unet which is supported is a standard conv net based unet. The temperature schedule is the same for Ising and for co. but I think that unbiased sampling is not possible at T = 0 as a target temperature because of the computation of 1/T. so it is not supported for the Ising model. In CO we never have terms that compute 1/T so it is not a problem there |
The problem is solved! |
I met another problem, when i run:
but i seems met cuda memory error: is there any method we can use multiple GPU? |
you can use multiple gpus by using '--GPUs 0 1 2 3' etc... |
sorry for asking problem once again, is there any possible to resume training from a checkpoint? |
yes, you can resume training with the usage of the script continue_training.py you just need to specify the GPUs and the wandb ID. For models trained with PPO there is a small bug that when training is resumed because the PPO uses a moving average of the average and std of the reward and I forgot to implement that those should also be stored and loaded. So for PPO resuming training does not work perfectly. But in most cases this should not be a problem. |
OK and thank you for detailed and patient explanation! |
It seems that U_net needs the value of "U_net_graph_dict" in lattice data, but during the generation of data, there is no computation of "U_net_graph_dict".
The text was updated successfully, but these errors were encountered: