Melody Huang's Avatar

Melody Huang

@melodyyhuang

Currently @ Yale, working on causal inference & cutting down on caffeine. Website: melodyyhuang.com

1,573
Followers
103
Following
13
Posts
20.09.2023
Joined
Posts Following

Latest posts by Melody Huang @melodyyhuang

Preview
Melody Huang (Yale) | Applied Statistics Workshop Gov 3009 Breadcrumbs

Weds at 12:00 ET: #Yale assistant professor @melodyyhuang.bsky.social presents "Relative Bias Under Imperfect Identification in Observational #CausalInference" at this week's #AppliedStatistics workshop. #politicalscience #statistics
appliedstatsworkshopgov3009.hsites.harvard.edu/event/melody...

04.03.2026 14:12 πŸ‘ 0 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

But have you seen Love Never Dies?? (It gets so much worse...)

24.02.2026 00:43 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Congrats to BART rider and Oakland legend Alysa Liu on winning a gold medal at the Olympics and making the Bay Area proud!

19.02.2026 22:21 πŸ‘ 7606 πŸ” 1074 πŸ’¬ 46 πŸ“Œ 132
The β€œproblem solving” approach and social science methodology – Cyrus Samii

I keep coming back to @cdsamii.bsky.social essay on the β€œproblem-centered” over β€œpuzzle-centered” research paradigm and I can’t help but feel like so many problems with social science methods boil down to this cyrussamii.com?p=3682

14.02.2026 20:02 πŸ‘ 26 πŸ” 7 πŸ’¬ 1 πŸ“Œ 0
Comic. Conjecture: It’s possible to construct a convincing proof without words, pictures, or content of any kind. Proof: [empty box] [caption] Proofs without words are cool, but we can go further.

Comic. Conjecture: It’s possible to construct a convincing proof without words, pictures, or content of any kind. Proof: [empty box] [caption] Proofs without words are cool, but we can go further.

Proof Without Content

xkcd.com/3201/

03.02.2026 20:56 πŸ‘ 2647 πŸ” 295 πŸ’¬ 40 πŸ“Œ 25

A blog post giving a more thorough take on survey experiments and the credibility revolution: cyrussamii.com?p=4168

03.12.2025 17:23 πŸ‘ 79 πŸ” 36 πŸ’¬ 6 πŸ“Œ 6

To me this is a depressing theme in modern academia.

There is so much work being produced, and so many competing demands on our time, that people rarely seem able to just closely read work and frankly say "yes, I believe this" or "no, I don't."

If we aren't doing this, what _are_ we doing?!

30.11.2025 10:57 πŸ‘ 46 πŸ” 9 πŸ’¬ 1 πŸ“Œ 2
Post image

new paper by Sean Westwood:

With current technology, it is impossible to tell whether survey respondents are real or bots. Among other things, makes it easy for bad actors to manipulate outcomes. No good news here for the future of online-based survey research

18.11.2025 19:15 πŸ‘ 777 πŸ” 390 πŸ’¬ 41 πŸ“Œ 126

Excellent report on experiences of @guidoimbens.bsky.social & Mary Wootters co-teaching "Causality, Decision Making, and Data Science" to undergrads at Stanford fall 2024: hdsr.mitpress.mit.edu/pub/uynpjlow... Course material here: stanford-causal-inference-class.github.io

03.11.2025 16:14 πŸ‘ 7 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Preview
Does AI help humans make better decisions? A statistical evaluation framework for experimental and observational studies | PNAS The use of AI, or more generally data-driven algorithms, has become ubiquitous in today’s society. Yet, in many cases and especially when stakes ar...

Excited to see our paper on evaluating whether AI can help humans make better decisions is now out in @pnas.org!
www.pnas.org/doi/10.1073/...

25.09.2025 02:41 πŸ‘ 10 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

Going to include this on my slides about partial identification

16.09.2025 19:39 πŸ‘ 12 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Population Studies Workshop | Institution for Social and Policy Studies

The Yale Population Studies Workshop is live! Check out our fall lineup here: isps.yale.edu/population-s... Speakers span sociology, economics, medicine, and political science. Both Yale and non-Yale folks are welcome to sign up for the email listserv!

26.08.2025 15:46 πŸ‘ 6 πŸ” 2 πŸ’¬ 0 πŸ“Œ 1
Preview
senseweight: Sensitivity Analysis for Weighted Estimators Provides tools to conduct interpretable sensitivity analyses for weighted estimators, introduced in Huang (2024) &lt;<a href="https://doi.org/10.1093%2Fjrsssa%2Fqnae012" target="_top">doi:10.1093/jrsssa/qnae012</a>&gt; and Hartman and Huang (2024) &lt;<a href="https://doi.org/10.1017%2Fpan.2023.12" target="_top">doi:10.1017/pan.2023.12</a>&gt;. The package allows researchers to generate the set of recommended sensitivity summaries to evaluate the sensitivity in their underlying weighting estimators to omitted moderators or confounders. The tools can be flexibly applied in causal inference settings (i.e., in external and internal validity contexts) or survey contexts.

New on CRAN: senseweight (0.0.1). View at https://CRAN.R-project.org/package=senseweight

22.08.2025 21:02 πŸ‘ 5 πŸ” 1 πŸ’¬ 0 πŸ“Œ 1

Could that be @melodyyhuang.bsky.social's new #rstats package for sensitivity analyses? πŸ‘€

22.08.2025 21:41 πŸ‘ 7 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0
Post image

Lots of different ways to do observational causal inferenceβ€”IV, proximal inference, etc. What if you could compare those strategies more directly?

New preprint w/ @melodyyhuang.bsky.social tries to do just that. Here's one cool figureβ€”we're able to visualize bias of 3 estimators on the same plot

01.08.2025 16:59 πŸ‘ 22 πŸ” 7 πŸ’¬ 3 πŸ“Œ 1

We show that the act of choosing an identification strategy implicitly expresses a belief about the degree of violations that must be present in alternative identification strategies. (4/4)

01.08.2025 14:17 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

To help compare the sensitivities across the different identification strategies, we propose a set of sensitivity tools and an augmented bias contour plot visualizes the relationship between these strategies. (3/n)

01.08.2025 14:17 πŸ‘ 0 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

We develop bias expressions for IV and proximal inference that show how violations of their respective assumptions are actually *amplified* by any unmeasured confounding in the outcome variable. (2/n)

01.08.2025 14:17 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Relative Bias Under Imperfect Identification in Observational Causal Inference To conduct causal inference in observational settings, researchers must rely on certain identifying assumptions. In practice, these assumptions are unlikely to hold exactly. This paper considers the bias of selection-on-observables, instrumental variables, and proximal inference estimates under violations of their identifying assumptions. We develop bias expressions for IV and proximal inference that show how violations of their respective assumptions are amplified by any unmeasured confounding in the outcome variable. We propose a set of sensitivity tools that quantify the sensitivity of different identification strategies, and an augmented bias contour plot visualizes the relationship between these strategies. We argue that the act of choosing an identification strategy implicitly expresses a belief about the degree of violations that must be present in alternative identification strategies. Even when researchers intend to conduct an IV or proximal analysis, a sensitivity analysis comparing different identification strategies can help to better understand the implications of each set of assumptions. Throughout, we compare the different approaches on a re-analysis of the impact of state surveillance on the incidence of protest in Communist Poland.

New working paper with @corymccartan.com on comparing the relative bias of different identification approaches! We compare the bias of selection-on-observables, instrumental variables, and proximal inference estimates under violations of their identifying assumptions.
arxiv.org/abs/2507.23743

01.08.2025 14:17 πŸ‘ 18 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0
4-panel comic. (1) [Person 1 with ponytail flanked by person with short hair and another person speaking into microphone at podium] PERSON 1: In the early 2010s, researchers found that many major scientific results couldn’t be reproduced. (2) PERSON 1: Over a decade into the replication crisis, we wanted to see if today’s studies have become more robust. (3) PERSON 1: Unfortunately, our replication analysis has found exactly the same problems that those 2010s researchers did. (4) [newspaper with image of speakers from previous panels] Headline: Replication Crisis Solved

4-panel comic. (1) [Person 1 with ponytail flanked by person with short hair and another person speaking into microphone at podium] PERSON 1: In the early 2010s, researchers found that many major scientific results couldn’t be reproduced. (2) PERSON 1: Over a decade into the replication crisis, we wanted to see if today’s studies have become more robust. (3) PERSON 1: Unfortunately, our replication analysis has found exactly the same problems that those 2010s researchers did. (4) [newspaper with image of speakers from previous panels] Headline: Replication Crisis Solved

Replication Crisis

xkcd.com/3117/

21.07.2025 23:54 πŸ‘ 4878 πŸ” 659 πŸ’¬ 28 πŸ“Œ 31
Post image

I have a new R package, 'bases,' out on CRAN today!

'bases' provides a number of basis expansions that you can use inside any modeling formula. This means you can fit nonparametric regressions with lm() or glmnet() easily!

Bases includes random Fourier features, approximate BART, and more!

29.05.2025 21:28 πŸ‘ 10 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0

We are pleased to announce an experiment intended to stimulate academic discussion and exchange, centered on papers published in Econometrica 1/5

27.03.2025 12:36 πŸ‘ 121 πŸ” 38 πŸ’¬ 2 πŸ“Œ 15
Screenshot of the linked Quarto website, with input checkboxes to change different conditions for a regression model that predicts economic performance based on US political party, with a reported p-value

Screenshot of the linked Quarto website, with input checkboxes to change different conditions for a regression model that predicts economic performance based on US political party, with a reported p-value

I’ve long used FiveThirtyEight’s interactive β€œHack Your Way To Scientific Glory” to illustrate the idea of p-hacking when I teach statistics. But ABC/Disney killed the site earlier this month :(

So I made my own with #rstats and Observable and #QuartoPub ! stats.andrewheiss.com/hack-your-way/

20.03.2025 18:30 πŸ‘ 1464 πŸ” 437 πŸ’¬ 57 πŸ“Œ 29

*Please repost* @sjgreenwood.bsky.social and I just launched a new personalized feed (*please pin*) that we hope will become a "must use" for #academicsky. The feed shows posts about papers filtered by *your* follower network. It's become my default Bluesky experience bsky.app/profile/pape...

10.03.2025 18:14 πŸ‘ 522 πŸ” 296 πŸ’¬ 23 πŸ“Œ 83

Does drinking this much coffee lower the risk of omitted variable bias?

11.03.2025 14:57 πŸ‘ 249 πŸ” 33 πŸ’¬ 10 πŸ“Œ 2

Congratulations Anton!!!

17.12.2024 16:45 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Congratulations Guilherme!

17.12.2024 16:11 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Key features:
- Closed-form solution: The robust CATE is an interpretable weighted average of site-specific CATE models.
- Flexibility: Allows the use of off-the-shelf single-site CATE estimation methods.
- Privacy-preserving: Avoids sharing individual-level data across sites. (4/n)

17.12.2024 16:11 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

We propose a minimax-regret framework for generalizing CATEs (conditional average treatment effects) across multisite data. We propose minimizing the worst-case regret over a class of target populations whose CATE can be represented as convex combinations of site-specific CATEs. (3/n)

17.12.2024 16:11 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Researchers often want to estimate HTEs that are consistent across different populations & contexts. When we have multiple source sites, we run into challenges:
(1) Site-specific models lack external validity.
(2) Pooled models risk bias if site heterogeneity is ignored. (2/n)

17.12.2024 16:11 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0