[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.5555/3524938.3524945guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
research-article
Free access

Efficient intervention design for causal discovery with latents

Published: 13 July 2020 Publication History

Abstract

We consider recovering a causal graph in presence of latent variables, where we seek to minimize the cost of interventions used in the recovery process. We consider two intervention cost models: (1) a linear cost model where the cost of an intervention on a subset of variables has a linear form, and (2) an identity cost model where the cost of an intervention is the same, regardless of what variables it is on, i.e., the goal is just to minimize the number of interventions. Under the linear cost model, we give an algorithm to identify the ancestral relations of the underlying causal graph, achieving within a 2-factor of the optimal intervention cost. This approximation factor can be improved to 1 + ε for any ε > 0 under some mild restrictions. Under the identity cost model, we bound the number of interventions needed to recover the entire causal graph, including the latent variables, using a parameterization of the causal graph through a special type of colliders. In particular, we introduce the notion of p-colliders, that are colliders between pair of nodes arising from a specific type of conditioning in the causal graph, and provide an upper bound on the number of interventions as a function of the maximum number of p-colliders between any two nodes in the causal graph.

Supplementary Material

Additional material (3524938.3524945_supp.pdf)
Supplemental material.

References

[1]
Adamic, L. A. and Huberman, B. A. Power-law distribution of the world wide web. Science, 287(5461):2115-2115, 2000.
[2]
Bareinboim, E. and Pearl, J. Causal inference and the data-fusion problem. Proceedings of the National Academy of Sciences, 113(27):7345-7352, 2016.
[3]
Bareinboim, E., Brito, C., and Pearl, J. Local characterizations of causal bayesian networks. In Graph Structures for Knowledge Representation and Reasoning, pp. 1-17. Springer, 2012.
[4]
Canonne, C. L., Diakonikolas, I., Kane, D. M., and Stewart, A. Testing conditional independence of discrete distributions. In 2018 Information Theory and Applications Workshop (ITA), pp. 1-57. IEEE, 2018.
[5]
Chan, S.-O., Diakonikolas, I., Valiant, P., and Valiant, G. Optimal algorithms for testing closeness of discrete distributions. In Proceedings of the twenty-fifth annual ACM SIAM symposium on Discrete algorithms, pp. 1193-1203. SIAM, 2014.
[6]
Eberhardt, F. Causation and intervention. PhD Thesis, Carnegie Mellon University, 2007.
[7]
Eberhardt, F. and Scheines, R. Interventions and causal inference. Philosophy of Science, 74(5):981-995, 2007.
[8]
Hauser, A. and Bühlmann, P. Characterization and greedy learning of interventional markov equivalence classes of directed acyclic graphs. Journal of Machine Learning Research, 13(Aug):2409-2464, 2012.
[9]
Hauser, A. and Bühlmann, P. Two optimal strategies for active learning of causal models from interventional data. International Journal of Approximate Reasoning, 55(4): 926-939, 2014.
[10]
He, Y.-B. and Geng, Z. Active learning of causal networks with intervention experiments and optimal designs. Journal of Machine Learning Research, 9(Nov):2523-2547, 2008.
[11]
Heinze-Deml, C., Maathuis, M. H., and Meinshausen, N. Causal structure learning. Annual Review of Statistics and Its Application, 5:371-391, 2018.
[12]
Hoyer, P. O., Janzing, D., Mooij, J. M., Peters, J., and Schölkopf, B. Nonlinear causal discovery with additive noise models. In Advances in neural information processing systems, pp. 689-696, 2009.
[13]
Hyttinen, A., Eberhardt, F., and Hoyer, P. O. Experiment selection for causal discovery. The Journal of Machine Learning Research, 14(1):3041-3071, 2013a.
[14]
Hyttinen, A., Hoyer, P. O., Eberhardt, F., and Jarvisalo, M. Discovering cyclic causal models with latent variables: A general sat-based procedure. arXiv preprint arXiv:1309.6836, 2013b.
[15]
Jukna, S. Extremal combinatorics: with applications in computer science. Springer Science & Business Media, 2011.
[16]
Katona, G. On separating systems of a finite set. Journal of Combinatorial Theory, 1(2):174-194, 1966.
[17]
Kisvölcsey, Á. Flattening antichains. Combinatorica, 1(26): 65-82, 2006.
[18]
Kocaoglu, M., Dimakis, A., and Vishwanath, S. Cost-optimal learning of causal graphs. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 1875-1884. JMLR. org, 2017a.
[19]
Kocaoglu, M., Shanmugam, K., and Bareinboim, E. Experimental design for learning causal graphs with latent variables. In Advances in Neural Information Processing Systems, pp. 7018-7028, 2017b.
[20]
Kruskal, J. B. The number of simplices in a complex. Mathematical Optimization Techniques, 10:251-278, 1963.
[21]
Lindgren, E., Kocaoglu, M., Dimakis, A. G., and Vishwanath, S. Experimental design for cost-aware learning of causal graphs. In Advances in Neural Information Processing Systems, pp. 5279-5289, 2018.
[22]
Loh, P.-L. and Bühlmann, P. High-dimensional learning of linear causal networks via inverse covariance estimation. The Journal of Machine Learning Research, 15(1):3065- 3105, 2014.
[23]
MacWilliams, F. J. and Sloane, N. J. A. The theory of error-correcting codes, volume 16. Elsevier, 1977.
[24]
Mao-Cheng, C. On separating systems of graphs. Discrete Mathematics, 49(1):15-20, 1984.
[25]
Parviainen, P. and Koivisto, M. Ancestor relations in the presence of unobserved variables. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 581-596. Springer, 2011.
[26]
Pearl, J. Causality: models, reasoning and inference, volume 29. Springer, 2000.
[27]
Pearl, J. Causality: Models, Reasoning, and Inference. Cambridge university press, 2009.
[28]
Richardson, T., Spirtes, P., et al. Ancestral graph markov models. The Annals of Statistics, 30(4):962-1030, 2002.
[29]
Shanmugam, K., Kocaoglu, M., Dimakis, A. G., and Vishwanath, S. Learning causal graphs with small interventions. In Advances in Neural Information Processing Systems, pp. 3195-3203, 2015.
[30]
Shimizu, S., Hoyer, P. O., Hyvärinen, A., and Kerminen, A. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(Oct): 2003-2030, 2006.
[31]
Shpitser, I. and Pearl, J. Identification of joint interventional distributions in recursive semi-markovian causal models. In Proceedings, The Twenty-First National Conference on Artificial Intelligence and the Eighteenth Innovative Applications of Artificial Intelligence Conference, July 16- 20, 2006, Boston, Massachusetts, USA, pp. 1219-1226, 2006.
[32]
Silva, R., Scheine, R., Glymour, C., and Spirtes, P. Learning the structure of linear latent variable models. Journal of Machine Learning Research, 7(Feb):191-246, 2006.
[33]
Spirtes, P., Glymour, C. N., Scheines, R., Heckerman, D., Meek, C., Cooper, G., and Richardson, T. Causation, prediction, and search. MIT press, 2000a.
[34]
Spirtes, P., Glymour, C. N., Scheines, R., Heckerman, D., Meek, C., Cooper, G., and Richardson, T. Causation, prediction, and search. MIT press, 2000b.
[35]
Tian, J. and Pearl, J. A general identification condition for causal effects. In AAAI, 2002.
[36]
Verma, T. and Pearl, J. An algorithm for deciding if a set of observed independencies has a causal explanation. In Uncertainty in artificial intelligence, pp. 323-330. Elsevier, 1992.
[37]
Zhang, K., Peters, J., Janzing, D., and Schölkopf, B. Kernel-based conditional independence test and application in causal discovery. In Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence, pp. 804-813. AUAI Press, 2011.

Cited By

View all
  • (2022)Generalizing goal-conditioned reinforcement learning with variational causal reasoningProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602194(26532-26548)Online publication date: 28-Nov-2022
  • (2022)Sample constrained treatment effect estimationProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3600661(5417-5430)Online publication date: 28-Nov-2022

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
ICML'20: Proceedings of the 37th International Conference on Machine Learning
July 2020
11702 pages

Publisher

JMLR.org

Publication History

Published: 13 July 2020

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)81
  • Downloads (Last 6 weeks)6
Reflects downloads up to 03 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Generalizing goal-conditioned reinforcement learning with variational causal reasoningProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602194(26532-26548)Online publication date: 28-Nov-2022
  • (2022)Sample constrained treatment effect estimationProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3600661(5417-5430)Online publication date: 28-Nov-2022

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media