NHD critique in Nature Reviews reviewed
I have been asked to pass comment on a review by Peter Kerr to be found in the pre-publication electronic release of Nature Reviews: Nephrology.
It goes by the title “Nocturnal haemodialysis – should we still provide this therapy?”
Peter is a particularly close personal friend, a much-valued colleague and a like-minded dialysis expert … though I fancy he is more the expert while I am more the evangelist – both for home haemodialysis. He is the head of one of Australia’s largest renal services and a staunch supporter and proponent of home-based care and extended hour and frequency home dialysis – as am I – and, in this country, that means nocturnal home haemodialysis. He oversees one of the largest nocturnal programs in our country.
I had actually been asked to write this review but was, at the time, on vacation in Canada and the request clearly passed to Peter. I am glad it did. He has written almost exactly as I would have written – though he has likely been more eloquent than I could have been – and, as such, I am thrilled to read his critique and agree with it, lock stock and barrel.
Here I am inserting a direct quote from Dori Schatell’s previous post (I hope Dori won’t mind) … as it encapsulates what I will then make more detailed comment on …[B]
[I]When the FHN trial was first started, the plan was to do two randomized, controlled studies. One compared IN-CENTER short daily (5-6 days a week) with IN-CENTER standard HD. This study was able to recruit enough people. Its results were resoundingly positive for each measure the researchers looked at.
The other study was intended to compare nocturnal home HD with standard in-center HD. But, there was a problem. Once people were educated about the options , they refused to be randomized to standard in-center HD! They wanted the better treatment. The study protocol had to be changed to nocturnal home HD vs. standard home HD (3x/week). Even then, the researchers were still not able to get enough people to take part. And, the differences between the two groups was not big enough to be statistically significant–though they did trend in the right direction.
The thing is, a paper by Saner has compared folks doing standard in-center HD with those doing standard HOME HD. Those on home HD lived about twice as long–even though they were doing the same amount of treatment. So, just the fact that both groups were at home means that there was less difference between them. And, I suspect that, once home, the “standard HD” folks–already educated about the benefits of MORE dialysis–may have dialyzed longer than they were supposed to for the study, or even did every other day. This would further reduce the differences between the groups and lead to a “negative” trial.[/I][/B]
The brief from Nature Reviews was to comment on the recent paper arising from the Frequent Hemodialysis Network (FHN) group, led by Dr Rocco, which compared the outcomes of frequent nocturnal haemodialysis with conventional haemodialysis. The trial was reported as a negative trial – but this was, in our view, rather misleading for a number of reasons. Nature Reviews – if my memory serves me correctly from the approach to me – set the title. There was no offered opportunity to add a ‘YES’ to the end of their pre-formed review title. In addition, the abstract does not convey any of the caveats to the FHN trial conclusions that many believe should have been given more emphasis. The body of Peter’s text nicely lists and discusses the many short-comings of the trial.
Statistics can be and often are misleading – even those produced from the hallowed ‘RCT’ … and my personal views on the certainty of the ‘RCT’ as the gold-standard for all trials and decisions that arise from trials can be found in a paper I wrote in NNI in 2008.
Some (myself included) believe that the frequent nocturnal arm of the FHN trial should not have been reported – certainly not as a ‘negative’ trial – and if reported, it should only have been regarded as a ‘mis-trial’ or as a ‘null-trial’. (And here, as an aside to Dori - maybe you could ask Carl Kjellstrand to comment on this comment!)
I should start out by saying that I should (and do) congratulate the FHN trial team for at least trying! They tried to do what has often been said couldn’t be done … doing an RCT of a major lifestyle intervention. BUT … the reason it has been so hard to do is that is almost impossible to take patients and randomise to an in-centre treatment (with shorter time, travel, institutionalisation … all the ‘goods’ and ‘bads’ of a centre program) or to home (with longer time, isolation, training time, home comforts … all the ‘goods’ and ‘bads’ of a home program) - and then expect the two to be comparable groups. This will always be the ‘flaw’ in life-style RCTs.
Dori correctly pointed out Saner’s paper - but the critics of this paper will always raise (correctly) the issue of ‘referral’ or ‘placement’ bias … that the ‘best’ patients go home and thus might be ‘expected’ to do better. This criticism of the results of home therapies is always hard to toss … and is true, to an extent. It muddies the water of every attempt to prove this is not the only reason for better home outcomes. Rocco’s effort with the FHN trial was laudable - but then failed to reach the power needed to ‘conclude’ - one way or the other though the ‘trend’ … and they were strong trends … was always to the NHD arm being the better outcome.
Sometimes its nice to mutter 'damn the statistics, I’ll follow my gut conclusions …", and just go home!
Now … from here on in, I will roughly paraphrase and/or partially quote Peter Kerr’s paper for you (in lay terms and as best I can) … I am sure he would not mind me doing so …
The trial set out to enroll 250 patients. The investigators had great trouble in reaching the numbers required to allow a likely statistical outcome to be achieved (in ‘stats’ terms, this is referred to as the power of a study).
The FHN trial of frequent (6x) NHD v conventional (3x) CHD trial was, in the end, quite severely and unfairly underpowered - and this applied to both ‘arms’ of the trial - the NHD arm and the CHD arm.
Only 87 patients in all could be recruited. Thsi is less than 35% of the number required to reach a statistical power to answer the 3 end-point questions of the trial … and these were to determine if was there a statistical difference between the NHD and CHD groups and did NHD show
- a difference in survival (death rate)
- a reduction in left ventricular hypertrophy (the wall thickness of the main pumping chamber of the heart as measured by the left ventricular mass)
- an improvement in physical well-being
Furthermore, at 12 months … a very short time to compare any dialysis-related outcome … of the already sadly underpowered 87 patients recruited, only 76 had data points for left ventricular mass!
As only 3 deaths occurred in the study period, it is very hard to make a statistical argument (positive or negative) from a very small number of deaths within a similarly small number of total patients. Yet a conclusion was made that as there was no statistical difference, that one therapy could not be supported above the other on the basis of survival … ie: it was reported as a negative trial for survival benefit for NHD.
Left ventricular mass did fall (this should be regarded a good outcome) – but it only reached a ‘p’ value – a level of significance – of 0.095. The confidence levels were -23.7 to +1.8.
What do these stats mean? … well, to be ‘significant’, the ‘p’ value should be less than 0.05 … or a more than 95% chance that the difference between the two groups is significant. A ‘p’ value of 0.095 means that there is only a 90.5% chance that the difference is significant. But, when the numbers in both groups are small, it is difficult to reach a significant ‘p’. That is what is mean by the ‘power’ of a study. The ‘power’ predicts the number fairly needed to detect significant differences between two groups. 250 patients split into two groups of 125 patients were needed, 76 split into two groups of less than 40 in each was achieved.
Let’s think further on ‘power’. If two patients are assigned, each to a different treatment, and the one assigned (A) feels better and the other assigned to (B) feels worse … is this proof that (A) is a better treatment than (B)?
No … this might just be by chance. You need lots of patients in each group to discern a ‘true’ benefit from a ‘chance’ benefit.
Now, assign 1000 patients, 500 to (A) and 500 to (B). Let’s say that in group (A) 490 feel better and 5 feel worse but in group (B), 450 feel worse and only 50 feel better. In this example, you might reasonable conclude – and applied stats would confirm – that (A) has a better-than-chance likelihood of being the superior treatment
A statistician would (fairly) say that this particular study – the FHN NHD vs CHD trial – was simply far too underpowered to show the statistical difference required to determine whether the two groups were truly different – or just different by chance.
The correct outcome, and more reasonable report - if indeed a report at all? … a ‘mis-trial’ and a ‘null’ report. In my view – and that of many others – that the results were reported at all, and then, as a negative study, was, correctly, incorrect!
Left ventricular mass reduction (a good thing) was found in the NHD group when compared to the CHD group and was measured as a 19.6% reduction. This was a similar dimension of LV mass reduction as had previously been shown in the other arm of the FHN trial (short daily HD) … where the reduction was reported as ‘significant’ - because of the larger number of patients.
A similar reduction - also reported as ‘positive’ or ‘significant’ was found in Bruce Culletons RCT from Calgary reported several years ago now and regarded (still) as a seminal trial - both in dialysis broadly and in NHD in particular.
The small ‘n’ (number of patients) again ‘muddied’ the conclusion in the FHN v CHD trial, leading to a ‘negative’ outcome where, not due to the of size of effect but due simply to the size of trial, it showed a ‘positive’ outcome in the (separate) report of the SDHD arm that was released at the ASN last year in Denver.
In addition, the amount of dialysis delivered also did not meet the required separation between the two groups.
The CHD group dialysed for quite a bit longer (in the US, the mean CHD run time is 213 minutes for CHD but the far-too-small CHD group in this reportedly ‘negative’ trial dialysed for a mean of 256 minutes).
The NHD group dialysed for quite a bit shorter than most NHD patients fro around the world are reported to commonly use (the mean 6x NHD run time more commonly 8hrs+ providing >40 hours per week while the also far-too-small NHD group in this reportedly ‘negative’ trial dialysed only for 6 hour runs and a mean 30 hours per week).
In addition, less than 75% (to be exact, 72.7% of NHD patients completed their allotted number of dialysis sessions.
As a result, there was less dialysis delivery difference between the two groups, already under powered by patient numbers, to attempt to prove a statistical difference between the amount of dialysis delivered.
In terms of the (always) soft and hard-to-determine physical composite health score results – the ‘do you feel better or worse’ questions – the NHD score showed a 7.4 point improvement against the CHD group.
Again … the trial reports that this ‘trended to’ but ‘did not reach’ significance. This is likely (again) a function of inadequate study power … ie: too few patients followed.
Better blood pressure control (I think quite important), and better phosphate control (also quite important) and both among the other ‘benefits’ that were seen in the NHD group, were not listed among the 3 primary endpoints and, as a result, didn’t not count in the assessment of ‘positivity’ or ‘negativity’ of the trial as a whole.
Peter’s critique reaches many other telling and correct conclusions beyond these few. His conclusion is 'that although the FHN trial reported by Rocco et al is a technically negative trial, it is unlikely to deter NHD (enthusiasm or) enthusiasts’. He concludes that “the trial may be dismissed (by the enthusiasts) as a negative trial, which is possibly the correct interpretation, but we should be encouraged that home haemodialysis is a safe and effective therapy.”
My own conclusion might have added to an otherwise thoughtful and spot-on critique - and well might I say “well done, PK - the much sadder thought that … … “the reporting of this study as ‘negative’ when, indeed, if probably better deserved to be either un-reported or described as ‘inadequately powered’ and ‘non-statistical’ is, in my view, regrettable– it is a truly ‘null study’ … and that it is “also regrettable, becuase it unfairly empowers the dialysis status quo” (short, unkind dialysis) while “it dis-empowers the impetus to better dialysis through longer and more frequent therapies”.
In my own view (and this is clearly a personal comment) it is a report which might better have never been published – and I am surprised it was, given the statistical weaknesses. It will set back what I beleive has been enlightened progress to a better dialysis future by entrenching the poor dialysis regimens of our past.