Optimal approach for training assuming limitless computing power and data #287
Replies: 2 comments
-
Hey @KristianBell , you know, this is a really difficult question to answer. Firstly, there is no one answer, or the answer is 'no, there is no ideal number of training files or set of settings'. There are however, some useful guidelines I can share...
I hope that helps. Good luck! |
Beta Was this translation helpful? Give feedback.
-
Many thanks for the info - much appreciated! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
What would the ideal settings and training file setup look like to create the most accurate recogniser, assuming limitless computing power?
The default settings for training (Epochs 100, Batch size 32, Learning rate 0.01) I assume are set as a compromise between fast performance and accuracy, but what would lead to better (less false positives and/or higher recall) results? Presumably a higher number of epochs and a larger batch size? But a smaller learning rate?
Similarly, is there an 'ideal' number of training files to use? Is it simply a case of the more the better? And is this the same with negative training files? Would it be useful for every species, to include as negative training files the training data for all other species? I can see how this would get computationally expensive but would it actually improve model precision, or would it make the model too conservative and reduce recall to almost nothing? Is there a 'best' negative training file type to use? For example, a sound most similar to, but not, the target species? Or is any other noise, either similar or otherwise, also useful as a negative training file?
I expect the answer to some extent depends on how important precision vs recall is to your particular use case but it would still be nice to know if bumping up epochs for example would be a 'safe' thing to do if you can hack the extra time, but I could see adding a huge number of negative training files may at some point work against model performance.
Beta Was this translation helpful? Give feedback.
All reactions