Skip to content

Releases: jpata/particleflow

v2.1.0

12 Nov 13:08
432cc49
Compare
Choose a tag to compare

What's Changed

Full Changelog: v2.0.0...v2.1.0

v2.0.0

22 Oct 12:23
fe4cbd5
Compare
Choose a tag to compare

What's Changed

  • add torch to requirements by @jpata in #353
  • Cleanup tensorflow and pyg by @jpata in #354
  • add compatibility table by @jpata in #355
  • CLIC dataset v2.3.0: fix target/truth momentum, st=1 to target more inclusive by @jpata in #352

Full Changelog: v1.9.0...v2.0.0

v1.9.0

07 Oct 07:04
3200518
Compare
Choose a tag to compare

What's Changed

  • fix CMS instructions by @jpata in #334
  • MLPF datasets v2.0.0: track pythia-level genjets, genmet in datasets; add per-particle ispu flag by @jpata in #332
  • CMS training instructions by @jpata in #336
  • Retraining with CMS samples v2.0.0 by @jpata in #337
  • Fix use of deprecated Ray Tune environment variable by @erwulff in #338
  • CMS dataset relabel, generate v2.1.0 with more stats, separate binary classifier by @jpata in #340
  • Pre-layernorm by @erwulff in #339
  • Switch to datasets v2.2.0 by @jpata in #341
  • try to improve val loss stability by @jpata in #342
  • Regression of log-transformed energy and pt, training checkpoints by @jpata in #343
  • CLIC dataset v2.2.0, CMS dataset 2.4.0 by @jpata in #345
  • Update dataset validation plot notebooks by @jpata in #347

Full Changelog: v1.8.0...v1.9.0

v1.8.0

13 Jun 10:15
98d59c2
Compare
Choose a tag to compare

What's Changed

The focus of this release is training on CMS datasets. The model has been retrained on high-statistics CMS samples and outperforms the baseline in our MLPF samples. The export of the transformer model to ONNX works with Flash Attention, and it can be integrated with CMSSW 14 and run on GPU. We have run the first physics validations in CMSSW, and we find that the performance with respect to the previous CMS version of MLPF is improved, but it does not yet outperform the baseline PF in CMSSW.

Some slides from the CMS progress:

The full list of PRs:

  • Remove pytorch geometric by @jpata in #310
  • add new paper to README by @jpata in #312
  • Add Ray Train training to GitHub actions CI/CD test by @erwulff in #314
  • CMSSW documentation by @jpata in #319
  • Full CMS pytorch training in May 2024 by @jpata in #316
  • update CMSSW validation scripts and documentation by @jpata in #322
  • onnx export with dynamic shapes, fast attention by @jpata in #324
  • switch ONNX model to full float for CMSSW compatibility by @jpata in #325
  • Update validation scripts to CMSSW_14_1_0 by @jpata in #323
  • update cmssw plots, add ttbar sample to valid, add multiparticlegun and vbf to training by @jpata in #330

Full Changelog: v1.7.0...v1.8.0

v1.7.0

08 Apr 06:09
4caf602
Compare
Choose a tag to compare

What's Changed

The primary feature of the new release is that pytorch is now the main mode of training.
The CMS status was presented at https://indico.cern.ch/event/1399688/#1-ml-for-pf.

  • switch pytorch training to tfds array-record datasets by @farakiko in #228
  • Timing the ONNX model, retrain CMS-GNNLSH-TF by @jpata in #229
  • fixes for pytorch, CMS t1tttt dataset, update response plots by @jpata in #232
  • fix pytorch multi-GPU training hang by @farakiko in #233
  • feat: specify number of samples as cmd line arg in pytorch training and testing by @erwulff in #237
  • Automatically name training dir in pytorch pipeline by @erwulff in #238
  • pytorch backend major update by @farakiko in #240
  • Update dist.barrier() and fix stale epochs for torch backend by @farakiko in #249
  • multi-bin loss in TF, plot fixes by @jpata in #234
  • PyTorch distributed num-workers>0 fix by @farakiko in #252
  • speedup of the pytorch GNN-LSH model by @jpata in #245
  • Implement HPO for PyTorch pipeline. by @erwulff in #246
  • fix tensorboard error by @farakiko in #254
  • fix config files by @erwulff in #255
  • making the 3d-padded models more efficient in pytorch by @jpata in #256
  • Fix pytorch inference after #256 by @jpata in #257
  • Update training.py by @jpata in #261
  • Reduce the number of data loader workers per dataset in pytorch by @farakiko in #262
  • fix inference by @farakiko in #264
  • Implementing configurable checkpointing. by @erwulff in #263
  • restore onnx export in pytorch by @jpata in #265
  • remove outdated forward_batch from pytorch by @jpata in #266
  • Separate multiparticlegun samples from singleparticle gun samples by @farakiko in #267
  • compare all three models in pytorch by @jpata in #268
  • Allows testing on a given --load-checkpoint by @farakiko in #269
  • added clic evaluation notebook by @jpata in #272
  • Fix --load-checkpoint bug by @farakiko in #270
  • Implement CometML logging to PyTorch training pipeline. by @erwulff in #273
  • Add command line argument to choose experiments dir in PyTorch training pipeline by @erwulff in #274
  • Implement multi-gpu training in HPO with Ray Tune and Ray Train by @erwulff in #277
  • Better CometML logging + Ray Train vs DDP comparison by @erwulff in #278
  • Fix checkpoint loading by @erwulff in #280
  • Learning rate schedules and Mamba layer by @erwulff in #282
  • use modern optimizer, revert multi-bin loss in TF by @jpata in #253
  • track individual particle loss components, speedup inference by @jpata in #284
  • Update the jet pt threshold to be the same as the PF paper by @farakiko in #283
  • towards v1.7: new CMS datasets, CLIC hit-based datasets, TF backward-compat optimizations by @jpata in #285
  • fix torch no grad by @jpata in #290
  • pytorch regression output layer configurability by @jpata in #291
  • Implement resume-from-checkpoint in HPO by @erwulff in #293
  • enable FlashAttention in pytorch, update to torch 2.2.0 by @jpata in #292
  • fix pad_power_of_two by @jpata in #296
  • Feat val freq by @erwulff in #298
  • normalize loss, reparametrize network by @jpata in #297
  • fix up configs by @jpata in #300
  • clean up loading by @jpata in #301
  • Fix unpacking for 3d padded batch, update plot style by @jpata in #306

Full Changelog: v1.6...v1.7.0

v1.6.2

04 Apr 20:14
cd2ed41
Compare
Choose a tag to compare

What's Changed

v1.6.1

04 Apr 20:14
Compare
Choose a tag to compare

What's Changed

v1.6

02 Oct 06:13
6f0fbce
Compare
Choose a tag to compare

What's Changed

  • pin matplotlib due to breaking changes in mpl 3.8.0 by @jpata in #210
  • Update README_tf.md by @jpata in #209
  • Update README.md by @jpata in #213
  • implement GNN-LSH model in torch by @jpata in #211
  • small fixes for CMS training, torch onnx export by @jpata in #215
  • updates for v1.6 by @jpata in #219
    • Improve the training and evaluation recipe for the benchmark model provided in the paper. Addresses #217
    • fix #222
    • fix #220
    • if padded batching is used, compute number of steps naively, do not step through the dataset to get the number of steps
    • split the Delphes dataset clearly to ttbar and qcd, like the other datasets
    • regenerate the Delphes dataset using tfds with array_record
    • fix #225

Full Changelog: https://github.com/jpata/particleflow/commits/v1.6

MLPF training with CLIC simulation

28 Aug 15:35
Compare
Choose a tag to compare
Pre-release

MLPF as in the upcoming paper, with training on CLIC simulation.
Cleanup of v1.5 with BFG to remove LFS errors.

What's Changed

New Contributors

Full Changelog: v1.4...v1.5

Baseline MLPF model for CMS

22 Sep 07:29
4cf744f
Compare
Choose a tag to compare