-
Notifications
You must be signed in to change notification settings - Fork 263
Description
Dear GRF community,
I understand that RF's and CF's can produce asymptotically valid standard errors and confidence intervals if certain rules are followed in their construction (e.g. honesty). I also understand that the (excellent!) GRF package generates standard errors as part of predictive output, if called.
My question: Could asymptotically valid s.e.'s also be obtained by bootstrapping the training sample, running the RF or CF (potentially with preceding residualization), generating predictions on the test sample (or a left-out observation from the training sample), and repeating many times? Then use the empirical standard deviation of the prediction as estimate for the standard error?
I'm asking because if bootstrapped s.e.'s are valid (= have the same desirable asymptotic properties as the ones generated by GRF), that would be very helpful in my current application, where I further process forest predictions using (highly nonlinear) functions.
Thank you very much for creating this user forum and for providing helpful feedback.