{"id":56,"date":"2009-07-06T04:00:31","date_gmt":"2009-07-06T11:00:31","guid":{"rendered":"http:\/\/www.mii.ucla.edu\/causality\/?p=66"},"modified":"2009-07-06T04:00:31","modified_gmt":"2009-07-06T11:00:31","slug":"resolving-disputes-between-j-pearl-and-d-rubin-on-causal-inference","status":"publish","type":"post","link":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/2009\/07\/06\/resolving-disputes-between-j-pearl-and-d-rubin-on-causal-inference\/","title":{"rendered":"On Myth, Confusion, and Science in Causal Analysis"},"content":{"rendered":"<p><strong><em>Andrew Gelman (Columbia) recently wrote a <a href=\"http:\/\/www.stat.columbia.edu\/~cook\/movabletype\/archives\/2009\/07\/disputes_about.html\">blog post<\/a>  motivated by Judea Pearl&#39;s paper, &quot;<a href=\"http:\/\/www.cs.ucla.edu\/~kaoru\/r348.pdf\">Myth, Confusion, and Science in Causal Analysis.<\/a> &quot; In response, Pearl writes:<\/em><\/strong><\/p>\n<p>Dear Andrew,<\/p>\n<p>Thank you for your blog post dated July 5.&nbsp;I appreciate your genuine and respectful quest to&nbsp;explore the differences between the approaches&nbsp;that I and Don Rubin are taking to causal inference.<\/p>\n<p>In general, I would be the first to rally behind your&nbsp;call for theoretical pluralism (e.g., &quot;It make sense that&nbsp;other theoretical perspectives such as Pearl&#39;s could be useful&nbsp;too.&quot;) We know that one can prove a theorem in geometry by either&nbsp;geometrical or algebraic methods, depending on the&nbsp;problem and the perspective one prefers to take&#8211;only the very dogmatic would label one of the&nbsp;methods &quot;unprincipled&quot;.<\/p>\n<p>My article, &quot;Myth, confusion and Science in Causal Analysis&quot;, is written with this dual perspective in mind, fully accommodating the graphical and potential-outcome conceptualizations as interchangeable, &quot;A theorem in one approach is a theorem in another,&quot; I wrote.<\/p>\n<p>However, when adherents of the one-perspective approach make claims that mathematically contradict those derived from the dual-perspective approach, one begins to wonder whether there is something more fundamental at play here.<\/p>\n<p><!--more--> In our case, the claims we hear from two adherents of the graph-less one-perspective school is: &quot;there is no reason to avoid adjustment for a variable describing subjects before treatment.&quot; And from three adherents of the graph-assisted dual-perspective school we hear: &quot;Adjustment for a variable &nbsp;describing subjects before treatment may be harmful.&quot;<\/p>\n<p>This is a blatant contradiction that affects every observational study and deserves therefore to be discussed even if we believe in &quot;let one thousand roses bloom.&quot;<\/p>\n<p>One may be tempted to resolve the contradiction by appealing to practical expediencies. For example,<\/p>\n<ol>\n<li>Nothing is black and white.<\/li>\n<li>Perhaps adjustment may be harmful in theory, but is very rare in practice,<\/li>\n<li>Perhaps the harm is really very small, or<\/li>\n<li>we do not really know in practice if it is harmful or not, so why worry?<\/li>\n<\/ol>\n<p>This line of defense would be agreeable, were it not accompanied with profound philosophical claims that the dual-perspective approach is in some way &quot;unprincipled&quot; and standing (God forbid) &quot;contrary to Bayesianism.&quot;<\/p>\n<p>The point is that we <em>DO KNOW<\/em> in practice when harm is likely to occur through improper adjustments. The same subjective knowledge that tells us that seat-belt usage does not cause smoking or lung disease also tells us that adjustment for seat-belt usage is likely to introduce bias.<\/p>\n<p>Moreover, one can derive this warning in the graph-less notation of potential outcome. So, the question remains: why haven&#39;t potential outcome scholars been issuing that warning to their students?<\/p>\n<p>The conjecture I made should concern every Bayesian and every educator, for it points beyond M-bias and covariate selection. The conjecture is that the language of &quot;potential outcome&quot; and &quot;ignorability&quot; discourages investigators from articulating and using valuable knowledge which they possess, for example, that seat-belt usage does not cause smoking. Do you know of any study where such a piece of knowledge was used in determining whether treatment assignment is &quot;ignorable&quot; or not? My conjecture is confirmed by potential-outcome practitioners who admit to be using &quot;ignorability&quot; invariably to justify their favorite method of analysis, never as an object to be justified by appeal to causal knowledge.<\/p>\n<p>As to indiscriminate conditioning in Bayesian philosophy, the example of controlling for an intermediate variable (between treatment and outcome) should illuminate our discussion. (I do not buy your statement that bias is &quot;tricky to define.&quot; It is extremely easy to define, even in Rubin&#39;s notation: &quot;Bias&quot; is what you get if you adjust for Z and treatment assignment is not ignorable conditioned on Z. This would suffice for our purposes)<\/p>\n<p>You say:<\/p>\n<ol>\n<li>A Bayesian analysis can control for intermediate outcomes&#8211;that&#39;s okay&#8211;but then&#8230;<\/li>\n<li>Jennifer and I recommend not controlling for intermediate outcomes.<\/li>\n<li>You can control for anything, you just then should suitable post process&#8230;.<\/li>\n<li>I heard Don Rubin make a similar point&#8230; Fisher made this mistake.<\/li>\n<\/ol>\n<p>Andrew, I know you did not mean it to sound so indecisive, but it does. Surely, one can always add 17.5 to any number, as long as one remembers to &quot;post-process&quot; and correct the mistake later on. But we are not dealing here with children arithmetic. Why not say it upfront: &quot;You cant arbitrarily add 17,5 to a number and hope you did not do any harm.&quot; Even the Mullahs of arithmetic addition would forgive us for saying it that way.<\/p>\n<p>If you incorporate an intermediate variable M as a predictor in your propensity score and continue to do matching as if it is just another evidentiary predictor, no post processing will ever help you, except of course, redoing the estimation afresh, with M removed.It will not fix itself by taking more samples. Is Bayesianism so dogmatic as to forbid us from speaking plainly and just say : &quot;Don&#39;t condition&quot;. (No wonder I once wrote: <a href=\"http:\/\/ftp.cs.ucla.edu\/pub\/stat_ser\/r284-reprint.pdf\">&quot;Why I am only a half-Bayesian&quot;<\/a>.)<\/p>\n<p>True, the great R A Fisher made a similar mistake. But it happened in the context of estimating &quot;direct effects&quot;, where one wants to control for the intermediary variable, not in the context of&quot;causal effects,&quot; where one wants the intermediaries to vary freely. Incidentally, the repair that Don Rubin offered in the Fisher lecture made things even worse. For example, the direct effect according to Rubin&#39;s definition (using principal stratification) is definable only in units absent of indirect effects. This means that a grandfather would be deemed to have no direct effect on his grandson&#39;s behavior in families where he has some effect on the father. In linear systems, to take a sharper example, the direct effect would be undefined whenever indirect paths exist from the cause to its effect.<\/p>\n<p>Such paradoxical conclusions emanating from a one-perspective culture underscore the wisdom, if not necessity of a dual-perspective analysis, in which the counterfactual notation Y<sub>x<\/sub>(u) is governed by the formal semantics of graphs, structural equations and open-mindedness.<\/p>\n<p>I just saw <a href=\"http:\/\/www.stat.columbia.edu\/~cook\/movabletype\/archives\/2009\/07\/disputes_about.html#comment-926476\">Larry Wasserman&#39;s comment<\/a>. Larry is right, I do not operate in an &quot;entirely different conceptual framework.&quot; I call the [X x Y] &#8211;&gt; [0,1] function P(Y<sub>x<\/sub> = y) &quot;causal effect&quot; leaving it up to the investigator to form differences P(Y<sub>1<\/sub> = y) &#8211; P(Y<sub>0<\/sub> = y), P(Y<sub>8<\/sub> = 3) &#8211; P(Y<sub>5<\/sub> = 3), or ratios: P(Y<sub>1<\/sub> = y) \/ P(Y<sub>0<\/sub> = y) or any other comparison that fits fashion and dogma.<\/p>\n<p>This does not make for a different conceptual framework, it is the common engineering practice of not wasting precious symbols on trivialities. What does call for a possible realignment of conceptual frameworks is what you tell your students about adjustment for intermediaries, and whether big-brother Bayes approves.<\/p>\n<p>Try it.<\/p>\n<p>Best,<\/p>\n<p>Judea<\/p>\n<p><em><strong><font color=\"#cc0000\">Update 1:<\/font> Please note that a parallel discussion is also underway on Gelman&#39;s blog. You may read the<br \/>\ncomments by <a href=\"http:\/\/www.stat.columbia.edu\/~cook\/movabletype\/archives\/2009\/07\/disputes_about.html#comments\">clicking here<\/a> .<\/strong><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Andrew Gelman (Columbia) recently wrote a blog post motivated by Judea Pearl&#39;s paper, &quot;Myth, Confusion, and Science in Causal Analysis. &quot; In response, Pearl writes: Dear Andrew, Thank you for your blog post dated July 5.&nbsp;I appreciate your genuine and respectful quest to&nbsp;explore the differences between the approaches&nbsp;that I and Don Rubin are taking to [&hellip;]<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[11,29],"tags":[],"class_list":["post-56","post","type-post","status-publish","format-standard","hentry","category-discussion","category-opinion"],"_links":{"self":[{"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/posts\/56","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/comments?post=56"}],"version-history":[{"count":0,"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/posts\/56\/revisions"}],"wp:attachment":[{"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/media?parent=56"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/categories?post=56"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/causality.cs.ucla.edu\/blog\/index.php\/wp-json\/wp\/v2\/tags?post=56"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}