Making the optimisation process more efficient: Pruning data and switching between acquisition functions. #1387
Replies: 3 comments 2 replies
-
For this see my comments here.
I can't give a definitive answer on this, but I have a strong feeling that rather than throwing data away in order to use a computationally expensive acquisition function you want to use that data to improve the model and instead use a cheaper acquisition function.
You can try reducing the number of fantasies here. This will reduce the accuracy of the approximation and it will depend on the problem how much of a hit optimization performance will take as a result. But using Are you using constraints? If so you can also pass a custom |
Beta Was this translation helpful? Give feedback.
-
I see, so if you're sampling repeatedly at a point what you'll do is refine the mean and variance estimates, that makes sense. In general that aggregation should ideally happen outside of GP model so that the number of data points stays reasonable and the model is fast enough to evaluate. This problems is also studied in https://arxiv.org/abs/1710.03206, which would be worth a read. |
Beta Was this translation helpful? Give feedback.
-
Say I am alternatingly using KG to get the global optimum and PM to look for optimum of the GP model. Although deleting data is not advised, I have been building the GP model with only the KG candidates that I tested. I have only used PM candidates to test if the optimum is performing better than before. I did this mainly to overcome optimisation wall-time and memory allocation issues, which were reccently addressed in #1704. Assuming no issues with adding more data in terms of the above bottlenecks (wall-time and memory), is it advisable to have as much data as possible (both PM and KG) to build the model? Will there be any issues with overfitting if I use I am working with 100-500 datapoints. |
Beta Was this translation helpful? Give feedback.
-
I am interested in using KG (Knowledge Gradient) method for optimisation, and I see that the process becomes too expensive beyond around 30 to 50 iterations. At the moment I am switching to EI (Expected Improvement) method when the KG method becomes expensive, but I want to continue using KG method. Towards this, I think pruning the data would help. I think datapoints that are too close by need to be removed. Also, I am not sure if this is already implemented or is it really worth it. So the following are my questions:
Beta Was this translation helpful? Give feedback.
All reactions