Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Identifying General Mechanism Shifts in Linear Causal Representations
by
Bello, Kevin
, Ravikumar, Pradeep
, Aragam, Bryon
, Locatello, Francesco
, Chen, Tianyu
in
Algorithms
/ Datasets
/ Nodes
/ Permutations
/ Representation learning
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Identifying General Mechanism Shifts in Linear Causal Representations
by
Bello, Kevin
, Ravikumar, Pradeep
, Aragam, Bryon
, Locatello, Francesco
, Chen, Tianyu
in
Algorithms
/ Datasets
/ Nodes
/ Permutations
/ Representation learning
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Identifying General Mechanism Shifts in Linear Causal Representations
Paper
Identifying General Mechanism Shifts in Linear Causal Representations
2024
Request Book From Autostore
and Choose the Collection Method
Overview
We consider the linear causal representation learning setting where we observe a linear mixing of \\(d\\) unknown latent factors, which follow a linear structural causal model. Recent work has shown that it is possible to recover the latent factors as well as the underlying structural causal model over them, up to permutation and scaling, provided that we have at least \\(d\\) environments, each of which corresponds to perfect interventions on a single latent node (factor). After this powerful result, a key open problem faced by the community has been to relax these conditions: allow for coarser than perfect single-node interventions, and allow for fewer than \\(d\\) of them, since the number of latent factors \\(d\\) could be very large. In this work, we consider precisely such a setting, where we allow a smaller than \\(d\\) number of environments, and also allow for very coarse interventions that can very coarsely \\textit{change the entire causal graph over the latent factors}. On the flip side, we relax what we wish to extract to simply the \\textit{list of nodes that have shifted between one or more environments}. We provide a surprising identifiability result that it is indeed possible, under some very mild standard assumptions, to identify the set of shifted nodes. Our identifiability proof moreover is a constructive one: we explicitly provide necessary and sufficient conditions for a node to be a shifted node, and show that we can check these conditions given observed data. Our algorithm lends itself very naturally to the sample setting where instead of just interventional distributions, we are provided datasets of samples from each of these distributions. We corroborate our results on both synthetic experiments as well as an interesting psychometric dataset. The code can be found at https://github.com/TianyuCodings/iLCS.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.