Skip to content

Bias in UncertaintyForest performance compared to paper #377

@EYezerets

Description

@EYezerets

My issue is about the fact that the UncertaintyForest benchmarks notebook shows that the UncertaintyForest class from ProgLearn underperforms IRF at d=20, which we did not see in the original paper.

I checked that samples are taken without replacement now in both the deprecated uncertainty-forest repo and in ProgLearn, i.e. bootstrap = False in the figure 2 tutorial in the uncertainty-forest repo, and replace = False in progressive-learner.py in ProgLearn. Also, I believe that the n_estimators (300), tree_construction_proportion (0.4), and kappa (3) values are the same.

Snapshot of documentation error:

From the paper (original Figure 2):
image

From benchmarks in EYezerets/ProgLearn on the fig2benchmark branch:

image

Additional context

Sorry, for some reason I'm not able to assign Richard to this issue. Could someone please help me include him in this conversation?

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions