EA - StrongMinds should not be a top-rated charity (yet) by Simon M
The Nonlinear Library: EA Forum - Un pódcast de The Nonlinear Fund
Categorías:
Link to original articleWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: StrongMinds should not be a top-rated charity (yet), published by Simon M on December 27, 2022 on The Effective Altruism Forum.GWWC lists StrongMinds as a “top-rated†charity. Their reason for doing so is because Founders Pledge has determined they are cost-effective in their report into mental health.I could say here, “and that report was written in 2019 - either they should update the report or remove the top rating†and we could all go home. In fact, most of what I’m about to say does consist of “the data really isn’t that clear yetâ€.I think the strongest statement I can make (which I doubt StrongMinds would disagree with) is:“StrongMinds have made limited effort to be quantitative in their self-evaluation, haven’t continued monitoring impact after intervention, haven’t done the research they once claimed they would. They have not been vetted sufficiently to be considered a top charity, and only one independent group has done the work to look into them.â€My key issues are:Survey data is notoriously noisy and the data here seems to be especially soThere are reasons to be especially doubtful about the accuracy of the survey data (StrongMinds have twice updated their level of uncertainty in their numbers due to SDB)One of the main models is (to my eyes) off by a factor of ~2 based on an unrealistic assumption about depression (medium confidence)StrongMinds haven’t continued to publish new data since their trials very early onStrongMinds seem to be somewhat deceptive about how they market themselves as “effective†(and EA are playing into that by holding them in such high esteem without scrutiny)What’s going on with the PHQ-9 scores?In their last four quarterly reports, StrongMinds have reported PHQ-9 reductions of: -13, -13, -13, -13. In their Phase II report, raw scores dropped by a similar amount:However, their Phase II analysis reports (emphasis theirs):As evidenced in Table 5, members in the treatment intervention group, on average, had a 4.5 point reduction in their total PHQ-9 Raw Score over the intervention period, as compared to the control populations. Further, there is also a significant visit effect when controlling for group membership. The PHQ-9 Raw Score decreased on average by 0.86 points for a participant for every two groups she attended. Both of these findings are statistically significant.Founders Pledge’s cost-effectivenes model uses the 4.5 reduction number in their model. (And further reduces this for reasons we’ll get into later).Based on Phase I and II surveys, it seems to me that a much more cost-effective intervention would be to go around surveying people. I’m not exactly sure what’s going on with the Phase I / Phase II data, but the best I can tell is in Phase I we had a ~7.5 vs ~5.1 PHQ-9 reduction from “being surveyed†vs “being part of the group†and in Phase II we had ~5.1 vs ~4.5 PHQ-9 reduction from “being surveyed†vs “being part of the groupâ€. For what it’s worth, I don’t believe this is likely the case, I think it’s just a strong sign that the survey mechanism being used is inadequate to determine what is going on.There are a number of potential reasons we might expect to see such large improvements in the mental health of the control group (as well as the treatment group).Mean-reversion - StrongMinds happens to sample people at a low ebb and so the progression of time leads their mental health to improve of its own accord“People in targeted communities often incorrectly believe that StrongMinds will provide them with cash or material goods and may therefore provide misleading responses when being diagnosed.†Potential participants fake their initial scores in order to get into the program (either because they (mistakenly) think there is some material benefit to being in the program or because they think it makes...
