Transportable Representations for Domain Generalization


AAAI2024


Kasra Jalaldoust, Elias Bareinboim
Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, 2024 Mar, pp. 12790-12800


technical report
Cite

Cite

APA   Click to copy
Jalaldoust, K., & Bareinboim, E. (2024). Transportable Representations for Domain Generalization. Proceedings of the AAAI Conference on Artificial Intelligence, 38, 12790–12800. https://doi.org/10.1609/aaai.v38i11.29175


Chicago/Turabian   Click to copy
Jalaldoust, Kasra, and Elias Bareinboim. “Transportable Representations for Domain Generalization.” Proceedings of the AAAI Conference on Artificial Intelligence 38 (March 2024): 12790–12800.


MLA   Click to copy
Jalaldoust, Kasra, and Elias Bareinboim. “Transportable Representations for Domain Generalization.” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, Mar. 2024, pp. 12790–800, doi:10.1609/aaai.v38i11.29175.


BibTeX   Click to copy

@article{kasra2024a,
  title = {Transportable Representations for Domain Generalization},
  year = {2024},
  month = mar,
  journal = {Proceedings of the AAAI Conference on Artificial Intelligence},
  pages = {12790-12800},
  volume = {38},
  doi = {10.1609/aaai.v38i11.29175},
  author = {Jalaldoust, Kasra and Bareinboim, Elias},
  month_numeric = {3}
}

Abstract

Generalizing across settings and changing conditions is one of the fundamental problems of AI. One critical assumption in this context is that the testing and training data come from the same distribution, which, despite its popularity in the literature, is often violated in practice. The anchors that allow generalizations to take place are causal and stem from the stability of the mechanisms underlying the system under investigation. Building on the theory of causal transportability introduced by Bareinboim & Pearl, we define the notion of ”transportable representations” to provide data structures for allowing a formal analysis of the domain generalization task. We then develop an algorithm to decide whether the distribution of the label, conditioned on the representation, can be computed in terms of the source distributions and assumptions about the commonalities and disparities across source and target domains. Moreover, we relax the assumption of having the graph as the task’s input and prove a graphical-invariance duality theorem, delineating the conditions under which certain invariances in the source data can be used as a sound and complete criterion for assessing the generalizability of a classifier. We review the prior literature and show how our findings provide a unifying theoretical perspective over several existing approaches to the domain generalization problem.