-
Notifications
You must be signed in to change notification settings - Fork 3
Open
Description
Hi team,
I’m trying to reproduce results and noticed some seeding gaps:
Sampling (test.py): The prior RNG is seeded via cfg.seed, but there’s no global torch.manual_seed / np.random.seed / random.seed, so other stochastic parts (e.g., model ops) are unseeded.
Training (liflow/experiment/train.py): DataModules pass cfg.seed into the torch.Generator and worker init for dataloaders (so shuffling/splits are deterministic), but the model initialization and any other randomness aren’t globally seeded. Also, priors default to seed=42 unless explicitly overridden in config, so they may diverge from cfg.seed.
Could you clarify:
- Should global seeds (torch, numpy, Python random) be set in training and sampling entrypoints for full reproducibility?
- Should prior seeds be tied to cfg.seed by default?
- Are there recommended deterministic settings (e.g., CUDA/Lightning flags) to match reported results?
Thanks!
Metadata
Metadata
Assignees
Labels
No labels