Skip to content

Bootstrapping standard errors for Random or Causal Forests? #1495

@kmoeltner

Description

@kmoeltner

Dear GRF community,
I understand that RF's and CF's can produce asymptotically valid standard errors and confidence intervals if certain rules are followed in their construction (e.g. honesty). I also understand that the (excellent!) GRF package generates standard errors as part of predictive output, if called.

My question: Could asymptotically valid s.e.'s also be obtained by bootstrapping the training sample, running the RF or CF (potentially with preceding residualization), generating predictions on the test sample (or a left-out observation from the training sample), and repeating many times? Then use the empirical standard deviation of the prediction as estimate for the standard error?

I'm asking because if bootstrapped s.e.'s are valid (= have the same desirable asymptotic properties as the ones generated by GRF), that would be very helpful in my current application, where I further process forest predictions using (highly nonlinear) functions.

Thank you very much for creating this user forum and for providing helpful feedback.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions