You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I am getting a strange issue whilist using the same script to train with cloudml. Last week I was training multiple models with the same script. Starting from a base model that gave pretty good results (acc = 0.9973; val_acc = 0.9925), I would add/remove elements or change parameters to find a preferred model architecture. All training runs were pretty much good, and nothing to be concerned.
Continuing my investigations today, the results are really bad. Restarted from the base model, but that also gave really bad results. Regardless of any changes to the model, the results are the same: after 1 epoch the accuracy flatlined at approx. 0.62. Training the same script on my linux machine, returns similar results to what I got last week. Changing the working directory do not help, but renaming the file seems to have worked (although the results were not as great, but acceptable).
Is this normal? Has this been noticed before? I think this is a cache problem, or is it the way I'm using cloudml? I really don't want to be using a different script when making small tweaks to the model.
Note: The data is saved to a local sub-directory named "data" in the working directory with the script, and not on Google storage bucket.
Hi, I am getting a strange issue whilist using the same script to train with cloudml. Last week I was training multiple models with the same script. Starting from a base model that gave pretty good results (acc = 0.9973; val_acc = 0.9925), I would add/remove elements or change parameters to find a preferred model architecture. All training runs were pretty much good, and nothing to be concerned.
Continuing my investigations today, the results are really bad. Restarted from the base model, but that also gave really bad results. Regardless of any changes to the model, the results are the same: after 1 epoch the accuracy flatlined at approx. 0.62. Training the same script on my linux machine, returns similar results to what I got last week. Changing the working directory do not help, but renaming the file seems to have worked (although the results were not as great, but acceptable).
Is this normal? Has this been noticed before? I think this is a cache problem, or is it the way I'm using cloudml? I really don't want to be using a different script when making small tweaks to the model.
Note: The data is saved to a local sub-directory named "data" in the working directory with the script, and not on Google storage bucket.
My keras/cloudml codes
The text was updated successfully, but these errors were encountered: