Pages

Monday, December 22, 2014

Increasing Placebo Responses in Psychiatric Drug Studies



In 2013, authors Rutherford and Roose  [American Journal of Psychiatry, 170  (7),  723-733] wrote a paper that discussed the results of a previous study that had found that the placebo (inactive "sugar" pill) response rates in random clinical trials (RCT's) of antidepressant medication had risen at a rate of 7% per decade over the past 30 years. Consequently, the average difference between active medication and placebo observed in published antidepressant trials decreased from an average of 6 points on the Hamilton Rating Scale for Depression (HAM-D) in 1982 to only 3 points in 2008.

Now the lead author of that paper and his colleagues have found something similar going on in RCT's between 1960 and 2013 of anti-psychotic medications for schizophrenia (the findings were published on line in October in JAMA Psychiatry). Most interestingly, in the 1960's, patients who received the placebo in such studies actually got worse on them.  By the 2000's, however, they were getting better on placebo.

Even more striking, the average RCT participant receiving an effective dose of medication in the 1960s improved by 13.8 points on the Brief Psychiatric Rating Scale (BPRS), whereas this difference diminished to 9.7 BPRS points by the 2000's.

What the heck is going on here? Are the medicines somehow becoming less effective than they used to be?

In their article from 2013, Rutherford et. al. try to explain this by looking at such things as expectancy (what do subjects think is going to happen with their symptoms), a statistical phenomenon known as  regression to the mean (see this post for a definition), the amount of contact subjects have with the study doctors, the social desirability of certain responses, and the “Hawthorne Effect” (subjects in an experiment improve or modify the aspect of their behavior under study simply by virtue of knowing that the behavior is being measured).

While the expectations of the average John Q. Citizen that antidepressants will work may have increased somewhat over the decades because of such things as celebrities describing their experiences with depression or commercials for Cymbalta and Abilify, there has also been a lot of negative information on that same score on television shows like Sixty Minutes and from the anti-psychiatry rants of Scientologists and others.

Whether these two influences completely cancel each other out is debatable, but I think it is safe to say that many of these possible reasons for a change in placebo response rate advanced by authors have in fact not changed significantly since the 1960's. In fact, if people in the 60's didn't think antidepressants would work, expectancies would have been lower, not only in the placebo group, but in the active treatment group as well.

If certain factors that affect placebo response rates have not changed a lot, then those factors can not explain the rise in the placebo response rates

The authors also mention a couple of factors that I believe to be more on the mark: first, that assessments of eligibility for a clinical trial may be biased toward inflated symptom reporting at the beginning of the study - when investigators have a financial incentive to recruit patients - and second, that most research participants in the 1960s and 1970s were recruited from inpatient psychiatric units, whereas current participants are symptomatic volunteers responding to advertisements.

The biggest change in research with RCT's over the period in question is that many studies are no longer done in medical schools, but by private entities called Contract Research Organizations (CRO's). The doctors who run the studies are paid for each subject they recruit, and subjects only get paid if they are recruited. This means that there are not just one set of financial incentives for everyone to exaggerate their symptoms at the beginning of a study, but two! This tendency will lead to a higher placebo response rate because after they are recruited, subjects no longer have an incentive to exaggerate their symptoms. So they seem to get better.

It is very easy to bias a research diagnostic interview. I'll get to that in a minute, but first a digression.

I was fortunate to train at a time when patients could be kept in the hospital for several months if necessary, so we got to see the patients in depth over a considerable time period, and could watch medication responses. People who have trained more recently do not see this any more.  Antidepressant responses clearly took a minimum of 2 weeks  - and then only if the patient responded to the very first drug given at the first dose given.

Because most patients do not understand this, the doctor can usually discriminate a placebo response from a true response by observing when the patient starts to get better combined with the rate at which they improve. Since subjects don't know what to expect, being on this timeline could not be due to the expectancy factor, which in turn is necessary for a having a good placebo response.

I can tell you that a severe, properly diagnosed melancholic depression almost never showed a significant placebo response.  The placebo response rate was probably about the same as the placebo response rate to a general anesthetic.

Another thing we observed was that patients with an acute schizophrenic reaction did not seem to get any better at all with such things as additional contact with doctors, which might be expected if a placebo response were taking place. In fact, the more you spoke with them, the more likely it would be that you would hear evidence that patients had a significant thought disorder than if you just had a briefer, casual conversation with them.

A thought disorder is at least as important as delusions and hallucinations in showing that someone is, in fact, someone with schizophrenia. People with a thought disorder see relationships between things that are completely illogical (loose associations). For example, the first patient I ever saw with schizophrenia in medical school believed that everyone who wore oxblood-colored shoes was a descendant of George Washington.  

Huh?

Anyway, back to the question of biasing diagnostic exams. This is particularly easy when diagnosing a clinical depression. It is important to distinguish them from those people who are merely chronically unhappy.  People with a clinical major depression, especially with so-called melancholic features, are a very different breed of cat.  

The  symptoms of both disorders do overlap a bit, so there are some cases in which it is really hard to tell one from the other. However, in the majority of cases it is a fairly easy call. It is, provided you do a complete psychiatric assessment, over several days, to see if a symptom of depression meets the requirement known as the Three P's: 

The symptoms need to be pervasive (they do not go away depending on what the patient is doing at a particular time), persistent (lasting almost all day every day for at least two weeks), and pathological (the patients symptoms and functioning differ to a highly significant degree from the patient's usual state). In addition to the three p's, all of the patient's symptoms have to always occur simultaneously.  

These types of characteristics do not usually show up on the type of symptom checklists used to assess patients in clinical trials, because the checklists are mostly based on a patient's self report.  Unfortunately, the majority of people do not know the difference between a clinically significant symptom and one that is not. The rate of false positive responses on checklists is staggering.

Many studies instead use something that is called a semi-structured diagnostic interview (such as one called a SCID) to make a diagnosis. It is called semi-structured because it tells the examiner to ask certain questions exactly as they are posed verbatim. However, the examiner is then free to ask any follow-up questions needed to clarify the clinical significance of any symptom the patient reports.

If you want to diagnose major depression regardless of whether or not the patient actually has it, all you have to do is accept every "yes" answer a patient gives to a question about a symptom without any follow-up questions to see if the symptom is characterized by the three P's. If the patient answers "No" to a question, however, you keep pumping the patient for additional clarification until you can find something the patient says that will justify changing the "no" answer to a "yes."  Voila.

No comments:

Post a Comment