This thesis investigates the integration of Explanation-Guided Learning (EGL) into causal inference, with a focus on Conditional Average Treatment Effect (CATE) estimation from observational data. While EGL has been widely studied in areas such as computer vision and natural language processing with an objective to improve explainability and generalizability of models, its application in the domain of causal inference remains largely unexplored. We propose a framework that incorporates EGL, specifically with counterfactual explanations, into modern CATE estimators: Meta-Learners. We improve accuracy of various meta-learning architectures by integrating explanation-based loss following the latest practices of the EGL literature. The proposed method enables the integration of structured domain knowledge into causal learning pipelines, which to the best of our knowledge, has not been done before. To evaluate the effectiveness of this approach, we conduct controlled experiments using synthetic data with known ground-truth treatment effects and varying levels of confounding. We assess performance across several meta-learners, including S-, X-, DR-, and domain adaptation variants. Our results demonstrate that incorporating explanation signals improves the accuracy of CATE estimation. In particular, the DA-learner and S-learner show the greatest reductions in mean absolute bias, with decreases of 49% and 51.8%, respectively. Overall, this work bridges two previously disconnected research areas – a newly emerged EGL and causal inference on observational data – and provides experimental proofs that demonstrate how explanation signals improve the accuracy of CATE estimation.

This thesis investigates the integration of Explanation-Guided Learning (EGL) into causal inference, with a focus on Conditional Average Treatment Effect (CATE) estimation from observational data. While EGL has been widely studied in areas such as computer vision and natural language processing with an objective to improve explainability and generalizability of models, its application in the domain of causal inference remains largely unexplored. We propose a framework that incorporates EGL, specifically with counterfactual explanations, into modern CATE estimators: Meta-Learners. We improve accuracy of various meta-learning architectures by integrating explanation-based loss following the latest practices of the EGL literature. The proposed method enables the integration of structured domain knowledge into causal learning pipelines, which to the best of our knowledge, has not been done before. To evaluate the effectiveness of this approach, we conduct controlled experiments using synthetic data with known ground-truth treatment effects and varying levels of confounding. We assess performance across several meta-learners, including S-, X-, DR-, and domain adaptation variants. Our results demonstrate that incorporating explanation signals improves the accuracy of CATE estimation. In particular, the DA-learner and S-learner show the greatest reductions in mean absolute bias, with decreases of 49% and 51.8%, respectively. Overall, this work bridges two previously disconnected research areas – a newly emerged EGL and causal inference on observational data – and provides experimental proofs that demonstrate how explanation signals improve the accuracy of CATE estimation.

Empirical Exploration of the Explanation-Guided Learning for Causal Effect Estimation

GLADO, ANNA
2025/2026

Abstract

This thesis investigates the integration of Explanation-Guided Learning (EGL) into causal inference, with a focus on Conditional Average Treatment Effect (CATE) estimation from observational data. While EGL has been widely studied in areas such as computer vision and natural language processing with an objective to improve explainability and generalizability of models, its application in the domain of causal inference remains largely unexplored. We propose a framework that incorporates EGL, specifically with counterfactual explanations, into modern CATE estimators: Meta-Learners. We improve accuracy of various meta-learning architectures by integrating explanation-based loss following the latest practices of the EGL literature. The proposed method enables the integration of structured domain knowledge into causal learning pipelines, which to the best of our knowledge, has not been done before. To evaluate the effectiveness of this approach, we conduct controlled experiments using synthetic data with known ground-truth treatment effects and varying levels of confounding. We assess performance across several meta-learners, including S-, X-, DR-, and domain adaptation variants. Our results demonstrate that incorporating explanation signals improves the accuracy of CATE estimation. In particular, the DA-learner and S-learner show the greatest reductions in mean absolute bias, with decreases of 49% and 51.8%, respectively. Overall, this work bridges two previously disconnected research areas – a newly emerged EGL and causal inference on observational data – and provides experimental proofs that demonstrate how explanation signals improve the accuracy of CATE estimation.
2025
Empirical Exploration of the Explanation-Guided Learning for Causal Effect Estimation
This thesis investigates the integration of Explanation-Guided Learning (EGL) into causal inference, with a focus on Conditional Average Treatment Effect (CATE) estimation from observational data. While EGL has been widely studied in areas such as computer vision and natural language processing with an objective to improve explainability and generalizability of models, its application in the domain of causal inference remains largely unexplored. We propose a framework that incorporates EGL, specifically with counterfactual explanations, into modern CATE estimators: Meta-Learners. We improve accuracy of various meta-learning architectures by integrating explanation-based loss following the latest practices of the EGL literature. The proposed method enables the integration of structured domain knowledge into causal learning pipelines, which to the best of our knowledge, has not been done before. To evaluate the effectiveness of this approach, we conduct controlled experiments using synthetic data with known ground-truth treatment effects and varying levels of confounding. We assess performance across several meta-learners, including S-, X-, DR-, and domain adaptation variants. Our results demonstrate that incorporating explanation signals improves the accuracy of CATE estimation. In particular, the DA-learner and S-learner show the greatest reductions in mean absolute bias, with decreases of 49% and 51.8%, respectively. Overall, this work bridges two previously disconnected research areas – a newly emerged EGL and causal inference on observational data – and provides experimental proofs that demonstrate how explanation signals improve the accuracy of CATE estimation.
Causal Inference
Explainable AI (xAI)
Explanation Guided
(CATE)
Neural Network-based
File in questo prodotto:
File Dimensione Formato  
Explanation-Guided Learning for Causal Inference.pdf

accesso aperto

Dimensione 2.53 MB
Formato Adobe PDF
2.53 MB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/108227