Skip to content

Commit 12ce810

Browse files
committed
Update notes
1 parent a47e861 commit 12ce810

File tree

3 files changed

+50
-0
lines changed

3 files changed

+50
-0
lines changed

notes/Master.bib

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,21 @@
1+
@article{cattaneo2025honest,
2+
title={The honest truth about causal trees: Accuracy limits for heterogeneous treatment effect estimation},
3+
author={Cattaneo, Matias D and Klusowski, Jason M and Yu, Ruiqi Rae},
4+
journal={arXiv preprint arXiv:2509.11381},
5+
year={2025}
6+
}
7+
8+
@article{athey2016recursive,
9+
title={Recursive partitioning for heterogeneous causal effects},
10+
author={Athey, Susan and Imbens, Guido},
11+
journal={Proceedings of the National Academy of Sciences},
12+
volume={113},
13+
number={27},
14+
pages={7353--7360},
15+
year={2016},
16+
publisher={National Academy of Sciences}
17+
}
18+
119
@article{agrawal2024automated,
220
title = {Automated efficient estimation using monte carlo efficient influence functions},
321
author = {Agrawal, Raj and Witty, Sam and Zane, Andy and Bingham, Elias},

notes/main.typ

Lines changed: 32 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -428,4 +428,36 @@ to measure the importance of variable $x$ by comparing the loss when including $
428428

429429
@wang2024multi proposed an extension of LOCO under multiple source data and using semiparametric theory to provide the inference of their measure.
430430

431+
432+
433+
== Ensemble learning
434+
435+
Ensemble learning, as the name suggests, combines multiple basde models to improve prediction performance. Common ensemble learning methods include bagging, boosting, and stacking.
436+
437+
=== Bagging
438+
BoostrapAggregating (bagging) is proposed by Leo Breiman @breiman1996bagging. The key idea is to generate multiple boostrap samples from the original data and train the base model on each boostrap sample then aggragate the predictions from all models, such as by averaging for regression or majority voting for classification.
439+
440+
==== Theory properties of bagging
441+
442+
- @breiman1996bagging Bagging can reduce the variance of unstable base models.
443+
- Peter Bühlmann and Bin Yu @buhlmann2002analyzing give some convergence rate analysis for bagging.
444+
- In the 2000s, many works on the theoretical understanding of bagging, seems no more work needed now.
445+
446+
==== Random forest
447+
448+
- Leo Breiman @breiman2001random proposed random forest, an ensemble learning method that builds multiple decision trees and merges their results to improve accuracy.
449+
450+
==== Causal forest
451+
452+
- @athey2016recursive proposed causal tree to estimate heterogeneous treatment effects namely the conditional average treatment effect (CATE) by extending decision tree, then @wager2018estimation using random forest to improve the estimation accuracy and provide asymptotic normality for inference.
453+
454+
- @cattaneo2025honest establishes an inconsistency lower bound on the point wise convergence rate of causal tree, and challenges the $alpha$-regularity condition (each split leaves at least a fraction $alpha$ of available samples on each side) needed to establish the convergence rate in @wager2018estimation.
455+
456+
=== Boosting
457+
458+
- XGBoost
459+
460+
=== Stacking
461+
462+
431463
#bibliography("Master.bib")

static/notes/notes.pdf

31 KB
Binary file not shown.

0 commit comments

Comments
 (0)