@@ -25,63 +25,53 @@ openbabel -- see installation instructions [here](http://openbabel.org/wiki/Cate
25
25
26
26
## Training
27
27
```
28
- usage: train.py [-h] -m MODEL -p PREFIX [-d DATA_ROOT] [-n FOLDNUMS] [-a]
29
- [-i ITERATIONS] [-s SEED] [-t TEST_INTERVAL] [-o OUTPREFIX]
30
- [-g GPU] [-c CONT] [-k] [-r] [--avg_rotations] [--keep_best]
31
- [--dynamic] [--cyclic] [--solver SOLVER] [--lr_policy LR_POLICY]
32
- [--step_reduce STEP_REDUCE] [--step_end STEP_END]
33
- [--step_when STEP_WHEN] [--base_lr BASE_LR]
34
- [--momentum MOMENTUM] [--weight_decay WEIGHT_DECAY]
35
- [--gamma GAMMA] [--power POWER] [--weights WEIGHTS]
36
- [-p2 PREFIX2] [-d2 DATA_ROOT2] [--data_ratio DATA_RATIO]
28
+ usage: train.py [-h] -m MODEL -p PREFIX [-d DATA_ROOT] [-n FOLDNUMS] [-a] [-i ITERATIONS] [-s SEED] [-t TEST_INTERVAL] [-o OUTPREFIX] [-g GPU] [-c CONT] [-k] [-r]
29
+ [--percent_reduced PERCENT_REDUCED] [--avg_rotations] [--checkpoint] [--keep_best] [--dynamic] [--cyclic] [--solver SOLVER] [--lr_policy LR_POLICY] [--step_reduce STEP_REDUCE]
30
+ [--step_end STEP_END] [--step_end_cnt STEP_END_CNT] [--step_when STEP_WHEN] [--base_lr BASE_LR] [--momentum MOMENTUM] [--weight_decay WEIGHT_DECAY] [--gamma GAMMA]
31
+ [--power POWER] [--weights WEIGHTS] [-p2 PREFIX2] [-d2 DATA_ROOT2] [--data_ratio DATA_RATIO] [--test_only] [--clip_gradients CLIP_GRADIENTS] [--skip_full]
32
+ [--display_iter DISPLAY_ITER] [--update_ratio UPDATE_RATIO]
37
33
38
34
Train neural net on .types data.
39
35
40
- optional arguments :
36
+ options :
41
37
-h, --help show this help message and exit
42
38
-m MODEL, --model MODEL
43
39
Model template. Must use TRAINFILE and TESTFILE
44
40
-p PREFIX, --prefix PREFIX
45
- Prefix for training/test files:
46
- <prefix>[train|test][num].types
41
+ Prefix for training/test files: <prefix>[train|test][num].types
47
42
-d DATA_ROOT, --data_root DATA_ROOT
48
43
Root folder for relative paths in train/test files
49
44
-n FOLDNUMS, --foldnums FOLDNUMS
50
- Fold numbers to run, default is '0,1,2'
51
- -a, --allfolds Train and test file with all data folds,
52
- <prefix>.types
45
+ Fold numbers to run, default is to determine using glob
46
+ -a, --allfolds Train and test file with all data folds, <prefix>.types
53
47
-i ITERATIONS, --iterations ITERATIONS
54
- Number of iterations to run,default 10 ,000
48
+ Number of iterations to run,default 250 ,000
55
49
-s SEED, --seed SEED Random seed, default 42
56
50
-t TEST_INTERVAL, --test_interval TEST_INTERVAL
57
- How frequently to test (iterations), default 40
51
+ How frequently to test (iterations), default 1000
58
52
-o OUTPREFIX, --outprefix OUTPREFIX
59
53
Prefix for output files, default <model>.<pid>
60
54
-g GPU, --gpu GPU Specify GPU to run on
61
- -c CONT, --cont CONT Continue a previous simulation from the provided
62
- iteration (snapshot must exist)
55
+ -c CONT, --cont CONT Continue a previous simulation from the provided iteration (snapshot must exist)
63
56
-k, --keep Don't delete prototxt files
64
- -r, --reduced Use a reduced file for model evaluation if exists(<pre
65
- fix>[_reducedtrain|_reducedtest][num].types)
66
- --avg_rotations Use the average of the testfile's 24 rotations in its
67
- evaluation results
57
+ -r, --reduced Use a reduced file for model evaluation if exists(<prefix>[reducedtrain|reducedtest][num].types). Incompatible with --percent_reduced
58
+ --percent_reduced PERCENT_REDUCED
59
+ Create a reduced set on the fly based on types file, using the given percentage: to use 10 percent pass 10. Range (0,100). Incompatible with --reduced
60
+ --avg_rotations Use the average of the testfile's 24 rotations in its evaluation results
61
+ --checkpoint Enable automatic checkpointing
68
62
--keep_best Store snapshots everytime test AUC improves
69
- --dynamic Attempt to adjust the base_lr in response to training
70
- progress
71
- --cyclic Vary base_lr between fixed values based on test
72
- iteration
63
+ --dynamic Attempt to adjust the base_lr in response to training progress, default True
64
+ --cyclic Vary base_lr in range of values: 0.015 to 0.001
73
65
--solver SOLVER Solver type. Default is SGD
74
66
--lr_policy LR_POLICY
75
- Learning policy to use. Default is inv .
67
+ Learning policy to use. Default is fixed .
76
68
--step_reduce STEP_REDUCE
77
- Reduce the learning rate by this factor with dynamic
78
- stepping, default 0.5
79
- --step_end STEP_END Terminate training if learning rate gets below this
80
- amount
69
+ Reduce the learning rate by this factor with dynamic stepping, default 0.1
70
+ --step_end STEP_END Terminate training if learning rate gets below this amount
71
+ --step_end_cnt STEP_END_CNT
72
+ Terminate training after this many lr reductions
81
73
--step_when STEP_WHEN
82
- Perform a dynamic step (reduce base_lr) when training
83
- has not improved after this many test iterations,
84
- default 10
74
+ Perform a dynamic step (reduce base_lr) when training has not improved after this many test iterations, default 5
85
75
--base_lr BASE_LR Initial learning rate, default 0.01
86
76
--momentum MOMENTUM Momentum parameters, default 0.9
87
77
--weight_decay WEIGHT_DECAY
@@ -90,14 +80,19 @@ optional arguments:
90
80
--power POWER Power, default 1
91
81
--weights WEIGHTS Set of weights to initialize the model with
92
82
-p2 PREFIX2, --prefix2 PREFIX2
93
- Second prefix for training/test files for combined
94
- training: <prefix>[train|test][num].types
83
+ Second prefix for training/test files for combined training: <prefix>[train|test][num].types
95
84
-d2 DATA_ROOT2, --data_root2 DATA_ROOT2
96
- Root folder for relative paths in second train/test
97
- files for combined training
85
+ Root folder for relative paths in second train/test files for combined training
98
86
--data_ratio DATA_RATIO
99
87
Ratio to combine training data from 2 sources
100
88
--test_only Don't train, just evaluate test nets once
89
+ --clip_gradients CLIP_GRADIENTS
90
+ Clip gradients threshold (default 10)
91
+ --skip_full Use reduced testset on final evaluation, requires passing --reduced
92
+ --display_iter DISPLAY_ITER
93
+ Print out network outputs every so many iterations
94
+ --update_ratio UPDATE_RATIO
95
+ Improvements during training need to be better than this ratio. IE (best-current)/best > update_ratio. Defaults to 0.001
101
96
```
102
97
103
98
MODEL is a caffe model file and is required. It should have a MolGridDataLayer
0 commit comments