-
Notifications
You must be signed in to change notification settings - Fork 31
Integrate wandb logging & sweep for ACTPolicy #28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
|
||
# Sweep entrypoint | ||
@classmethod | ||
def sweep_entrypoint(cls): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it necessary to define this method in a per-policy class?
If it is sufficient to define it in the if args.sweep:
block of Train.py, that would be simpler.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks! You're right — I’ve moved the sweep logic to Train.py's if args.sweep: block as suggested.
Thanks for the contribution! We're not currently using WandB, but if it becomes truly necessary, we'd be happy to merge it. (Before merging, we'd like to clean it up—for example, by moving as much of the code added to I'm curious whether sweeps are practically useful. Did using sweeps help you find better hyperparameters? |
Thank you for the explanation. However, the difficulty in imitation learning is that even if the validation loss is small, it does not necessarily mean that the task success rate is high at rollout of the policy. This is mentioned in Appendix G of https://arxiv.org/abs/2108.03298 |
Summary
This PR integrates Weights & Biases (wandb) into the ACTPolicy training pipeline of the RoboManipBaselines project, adding:
This is an initial integration targeting
TrainAct.py
and CLI entrypoint (Train.py
) for ACTPolicy.Changes
TrainAct.py
)Train.py
)Notes
wandb login
) before training, even if not using--sweep
pip install wandb
) if not already presentTrain.py
ACTPolicy
— other policies can be added in the futurewandb/local
Docker solution for offline/local logging, but encountered unresolved errors — currently defaulting to online (cloud-based) usage.How to test
Run normal training for ACTPolicy:
Run sweep training for ACTPolicy:
Verify:
output/ckpt/
Results:




Checklist
Future work
Mlp
,Sarnn
,MtAct
,DiffusionPolicy
)Known limitations
ACTPolicy