Development of Prognosis in Palliative care Study (PiPS) predictor models to improve prognostication in advanced cancer: prospective cohort studyBMJ 2011; 343 doi: http://dx.doi.org/10.1136/bmj.d4920 (Published 25 August 2011) Cite this as: BMJ 2011;343:d4920
- Bridget Gwilliam, clinical research fellow1,
- Vaughan Keeley, consultant in palliative medicine2,
- Chris Todd, professor of primary care and community health and director of research3,
- Matthew Gittins, statistician4,
- Chris Roberts, reader in biostatistics4,
- Laura Kelly, Macmillan consultant in palliative medicine5,
- Stephen Barclay, Macmillan post-doctoral research fellow6,
- Patrick C Stone, reader in palliative medicine1
- 1Division of Population, Health Sciences and Education, St George’s University of London, London SW17 0RE, UK
- 2Royal Derby Hospital, Derby, UK
- 3School of Nursing, Midwifery and Social Work, University of Manchester, Manchester, UK
- 4Health Sciences, School of Community Based Medicine, University of Manchester
- 5Macmillan Consultant in Palliative Care Team, East Surrey Hospital, Surrey and Sussex Healthcare NHS Trust, Redhill, Surrey, UK
- 6General Practice and Primary Care Research Unit, Department of Public Health and Primary Care, Institute of Public Health, Cambridge, UK
- Correspondence to: P C Stone
- Accepted 4 July 2011
Objective To develop a novel prognostic indicator for use in patients with advanced cancer that is significantly better than clinicians’ estimates of survival.
Design Prospective multicentre observational cohort study.
Setting 18 palliative care services in the UK (including hospices, hospital support teams, and community teams).
Participants 1018 patients with locally advanced or metastatic cancer, no longer being treated for cancer, and recently referred to palliative care services.
Main outcome measures Performance of a composite model to predict whether patients were likely to survive for “days” (0-13 days), “weeks” (14-55 days), or “months+” (>55 days), compared with actual survival and clinicians’ predictions.
Results On multivariate analysis, 11 core variables (pulse rate, general health status, mental test score, performance status, presence of anorexia, presence of any site of metastatic disease, presence of liver metastases, C reactive protein, white blood count, platelet count, and urea) independently predicted both two week and two month survival. Four variables had prognostic significance only for two week survival (dyspnoea, dysphagia, bone metastases, and alanine transaminase), and eight variables had prognostic significance only for two month survival (primary breast cancer, male genital cancer, tiredness, loss of weight, lymphocyte count, neutrophil count, alkaline phosphatase, and albumin). Separate prognostic models were created for patients without (PiPS-A) or with (PiPS-B) blood results. The area under the curve for all models varied between 0.79 and 0.86. Absolute agreement between actual survival and PiPS predictions was 57.3% (after correction for over-optimism). The median survival across the PiPS-A categories was 5, 33, and 92 days and survival across PiPS-B categories was 7, 32, and 100.5 days. All models performed as well as, or better than, clinicians’ estimates of survival.
Conclusions In patients with advanced cancer no longer being treated, a combination of clinical and laboratory variables can reliably predict two week and two month survival.
Patients with advanced cancer and their carers often wish to know how long they have left to live.1 2 Accurate prognostic information can allow patients adequate time to prepare for their impending death.3 Qualitative studies show that patients in palliative care want to be given honest and accurate prognostic information but that this information needs to be shared sensitively and in a way that respects patients’ desire to maintain hope.4 5
Prognostic information is also important for clinicians. Realistic survival estimates can inform decisions about the appropriateness of medical interventions and the timing of referral to specialist palliative care services or admission to a hospice. Clinicians’ predictions are routinely used to prioritise patients who are suitable for inclusion in programmes such as the Gold Standards Framework,6 to determine which patients are suitable for “fast-tracking” arrangements for referral to community care, and to determine eligibility for clinical trials.
Clinicians’ predictions of survival are inaccurate and over-optimistic.7 Moreover, their estimates are affected by factors such as training, experience, seniority, and level of acquaintance with the patient.8 For these reasons, a clinician’s prediction is not a very reliable or robust method of predicting survival. None the less, clinicians’ predictions are widely used, and any attempt to produce a more standardised answer to the question “How long have I got?” should be able to show that it is at least as reliable as this approach.
Previous studies have identified several clinical and laboratory variables that predict survival in patients with advanced cancer.9 10 Prognostic tools have been created by using scoring systems derived from combinations of these variables.11 Some of these tools include clinicians’ subjective estimates.12 13 Some rely on observers’ ratings of clinical variables but do not consider the added value of including laboratory data.14 15 Some tools include laboratory data but consequently can be applied only when such data are available.13 16 17 Some tools were developed by using only competent patients,13 16 17 whereas confused patients were included in the process of scale development of other tools.14 18 One important criticism of existing tools is that their performance has not been “benchmarked” against clinicians’ predictions of survival, making it difficult to judge the usefulness of these scales in clinical practice.
We sought to develop a prognostic tool that could be easily applied in clinical practice. Patients with advanced cancer are often frail, vulnerable, or confused. Many patients are unwilling to undergo further procedures (even blood tests). To produce a practical prognostic tool, we aimed to create a composite scoring system that could be used in both competent and incompetent patients and regardless of whether laboratory data were available. We wanted to develop a scoring system that was applicable across a range of palliative care settings (hospice, hospital, and community). Moreover, we wanted to produce a scale that did not rely on clinicians’ estimates of survival but was at least as accurate as their best predictions.
This was a multicentre study involving 18 palliative care services across England. Participating units included hospital support teams, hospice inpatient units, day centres, and community palliative care teams. We collected data between March 2006 and August 2009 and followed up all patients for a minimum of three months after recruitment.
Patients were eligible to participate in this study if they had been newly referred to the relevant palliative care service with a diagnosis of advanced (locally extensive or metastatic), incurable cancer. Eligible patients were no longer receiving active treatment for cancer, and no further disease modifying treatment was planned. Patients with hormone sensitive tumours who were still receiving hormone therapy but who had developed hormone resistant disease were eligible to participate, as were patients receiving palliative radiotherapy that was not expected to prolong survival (for example, for bone metastases). Both competent and incompetent patients were eligible. We restricted study assessments in incompetent patients to observers’ ratings of clinical status and extraction of information from case notes (questionnaires and blood tests were not done in incompetent patients).
Systematic reviews of studies involving patients with advanced cancer identified several variables with good a priori evidence of prognostic utility.9 10 These variables formed the basis for our assessments in study participants.
Observer rated symptom checklist, performance status, and global health status
In consultation with the clinical team, a researcher completed a checklist of the following symptoms: pain, breathlessness at rest, loss of appetite, dry mouth, difficulty swallowing, and tiredness. We recorded these symptoms as being either present or absent. We recorded the patient’s Eastern Co-operative Oncology Group (ECOG) performance status.19 ECOG scores vary between 0 and 4 (0=normal functional abilities, 4=confined to a bed or chair and requires all care). We recorded global health status by using a study specific seven point scale (1=extremely poor health, 7=normal health).
We measured weight and height when possible and asked patients or their carers whether weight loss had occurred over the previous month. We recorded pulse rate and the presence of clinically apparent dependent pitting oedema or ascites.
Abbreviated mental test score
We used the abbreviated mental test score to assess cognitive status.20 Although not a comprehensive assessment of cognitive function, this 10 item scale is often used in clinical practice to screen patients for confusion. We chose it for its ease of administration and because it reflects the type of assessment commonly done in clinical practice. For ethical reasons, patients who were deemed to be incompetent were not interviewed for this study but were attributed a score of 0.
Clinician’s prediction of survival
We asked clinicians to predict the survival of the patient by using one of four categories: “days” (that is, less than 14 days), “weeks” (from two weeks to less than eight weeks), “months” (from two months to less than 12 months), and “years” (12 months or more). We chose these categories as they seemed to have the greatest face validity among palliative care practitioners. Clinicians seldom make precise predictions, and asking clinicians to predict survival to the nearest day or week would have resulted in spuriously accurate prognostic predictions. Moreover, the Department of Health’s Prognostic Indicator Guidance, issued as part of the Gold Standards Framework,6 recommends these categories. Researchers obtained a clinician’s estimate independently from both a doctor and a nurse. When these estimates agreed, we also took this as the “agreed multi-professional estimate.” When a discrepancy existed between the clinicians, we asked them to discuss the case and arrive at an explicitly agreed estimate. Because only a small number of participants were estimated to be likely to survive for “years,” we included these patients with those with an estimated survival of “months” in all analyses.
Demographic, disease related, and treatment related variables
We recorded the age, sex, ethnicity, marital status, normal residence, diagnosis, and extent of disease (site of primary tumour and presence and sites of metastatic disease). We also recorded information on the extent of previous treatments, the time since the original diagnosis, and the time since the last treatment for cancer had been administered. We collected data on survival from entry to the study by flagging all patients for mortality with the NHS Information Centre.
We recorded comorbidity by using the Charlson Co-morbidity Index.21 This provides a measure of the overall extent of comorbid disease by using a four point scoring system (0=no comorbid diseases, 3=presence of multiple or severe comorbidity).
Laboratory variables (when available)
We asked all competent patients who agreed to participate in the study to provide a specimen of blood for analysis of haematological and biochemical parameters. We did not require incompetent patients to provide a blood specimen for analysis; however, where such results were available from clinical practice (within four days of study entry), we entered them into the study database and included them in the analysis.
We asked competent patients to complete the symptom checklist, ECOG performance status, and global health status. We also asked competent patients whether they wished to estimate their own prognosis. We did not use these data in the generation of the prognostic scores reported here, and they will be reported elsewhere.
We maintained a screening log of all clinical referrals to participating units. We identified eligible competent patients and, with the agreement of the clinical team, provided them with information and invited them to participate. We similarly informed the relatives or carers of eligible incompetent patients and asked them for assent.
We assessed study participants on two occasions one week apart: blood tests were not repeated specifically for the study, as this was judged to be too onerous for participants. We used only data from the baseline assessment to construct the prognostic tools described in this paper, and data from the second assessment are not presented. We calculated survival (in days) from the date of study entry.
We constructed a database and checked data for accuracy and missing values. We included in the model building analysis those variables that we deemed a priori to be predictors of survival on the basis of clinical knowledge and for which only small amounts of data were missing (<6% of data missing for all included variables except C reactive protein, for which 13% of data were missing). We included participants with incomplete data in the analysis by using multiple imputation by chained equations,22 23 with 20 imputations. No consensus exists about the best method for selecting variables for inclusion in a predictive tool, but backwards elimination is generally the preferred method, and we chose it for this analysis.24 To reduce the computational burden, we used single imputation for the initial stages of backward selection with a cut-off P value of 0.3 before completing the selection of variables by using multiple imputation with a cut-off P value of 0.05. We did this analysis for outcomes at two weeks’ (14 day) and two months’ (56 day) survival in both the full dataset (PiPS model A) and the restricted dataset obtained from participants for whom blood results were available (PiPS model B), giving four models in all. An alternative modelling strategy would have used a single ordinal logistic model, but the data did not satisfy the proportional odds assumption, evident from the difference between the two week and two month models for both PiPS models A and B. We estimated the level of over-optimism in each model by using a bootstrap technique,22 25 26 27 in which we drew bootstrap samples from the original dataset and then repeated the complete model selection procedure to develop a new model. We then estimated optimism as the difference in performance between the bootstrap sample and the original dataset. We used the average of the optimism over all 1000 bootstraps as a correction factor for the performance of the original model based on the full dataset.
We assessed the goodness of fit of the models by using the area under the receiver operating characteristics curve. We combined the models’ predictions for the two week and two month cut-off points to produce a categorical prediction of survival (“days,” “weeks,” or “months/years”) that could be directly compared with clinician’s estimates. One limitation of fitting two logistic models is that some participants may be classified as having a survival of both less than two weeks and more than two months. This happened on only one occasion, and we classified this case as having a survival between two weeks and two months. We used linear weighted κ to compare the performance of the clinicians with that of the models (after adjustment for over-optimism), as this averages the κ coefficients for the two week and two month threshold.28 We plotted Kaplan-Meier survival curves for each of the three risk groups (“days,” “weeks,” and “months+”) and used Harrel’s C statistic to measure concordance.29
During the study period, 7017 consecutive eligible patients were identified across the 18 participating palliative care services. The research team was able to access only 2401 of the eligible patients (fig 1⇓). The most common reason for failure to access patients was discharge or death before being approached by the research team. The clinical team denied the researchers access to eligible patients on 1102 occasions, because they judged that involvement in a research study would prove to be too distressing for patients or their carers. Of the 2401 patients/carers who were approached by the research team, 1023 (43%) agreed to participate (780 competent, 243 incompetent patients). No significant differences in age, sex, or distribution of diagnoses existed between recruited patients and the eligible patients who were not recruited. Five patients were removed from the dataset: one was subsequently diagnosed as having a new primary tumour and no longer met the eligibility criteria, two chose to withdraw, and two were recruited twice in error. The final sample thus consisted of 1018 participants. Tables 1⇓ and 2⇓ show the characteristics of the study sample. The median survival of the group was 34 days.
Development of PiPS-A and PiPS-B models
We developed the PiPS-A models by using the core dataset that was collected from all participants (both competent and incompetent). We developed separate models to predict survival of two weeks (14 days) or more (PiPS-A14) and to predict survival of two months (56 days) or more (PiPS-A56). We developed the PiPS-B models by using data from only those participants for whom blood results had also been obtained. Again, we developed separate models for predicting two week (PiPS-B14) and two month (PiPS-B56) survival. Tables 3⇓ and 4⇓ show the results of the backward stepwise logistic regression.
Performance of models and cross validation
The areas under the curve of the four models all exceeded 0.79 (tables 3⇑ and 4⇑), which shows good discrimination.24 Logistic regression models provide estimates of survival at specified “cut-off points.” However, clinicians are not usually interested in just whether a patient will survive for more than, for example, two weeks (the information provided by PiPS-A14). For this reason, the clinical question is often framed in terms of whether the patient is expected to survive for “days” (<14 days), “weeks” (14-56 days), or “months/years” (>56 days). To consider this question, we combined the PiPS-A14 and PiPS-A56 models (and the PiPS-B14 and PiPS-B56 models) to predict whether a patient was likely to survive for more than two weeks but less than two months. We then compared these predictions with the multi-professional estimates of survival (tables 5⇓ and 6⇓). Using this more demanding measure of clinical utility, the PiPS-A models performed at least as well as the clinicians (PiPS-A predictions correct on 59.6% of occasions and multi-professional predictions correct on 57.5%). The PiPS-B models performed significantly better than did either the doctors (61.5% v 52.6%; P=0.0135) or the nurses (61.5% v 52.3%; P=0.012) but were not significantly better than the multi-professional estimate (61.5% v 53.7%; P=0.188).
Using the PiPS-A models (fig 2⇓), the median survival of patients predicted to survive “days” was five days, that of those predicted to survive “weeks” was 33 days, and that of those predicted to survive “months+” was 92 days (Harrell’s concordance (C) statistic=0.69). For the PiPS-B models (fig 3⇓), the equivalent median survival was seven days, 32 days, and 100.5 days (C=0.67).
In this large, prospective, multicentre study, we have developed and validated four prognostic models for predicting survival in palliative care patients with advanced cancer. These models are able to identify reliably those patients with expected prognoses of “days,” “weeks,” or “months/years” and can be used in either competent or incompetent patients and in circumstances when blood results are available and when additional investigations would be inappropriate. When combined, the prognostic models were at least as good as a multi-professional clinical estimate of survival; when blood results are available, the models were significantly better than either a doctor’s or a nurse’s prediction (but not a multi-professional estimate).
Strengths of study
Our study had several strengths. The primary aim was to develop a prognostic scoring system that was significantly better than (and independent of) clinicians’ predictions of survival. No previous studies have attempted to “benchmark” their performance against current best clinical practice in this way. As a result, we have been able to show that the PiPS scores are better than the best uni-professional specialist clinical predictions of survival. All of the variables assessed as part of the PiPS study were previously identified in systematic reviews as being of likely prognostic significance,9 10 and the final variables included in our models have been found to be reliable predictors of survival in a large heterogeneous group of patients with advanced cancer in a variety of different settings (hospital, hospice, and community). This supports the generalisability of our results. Our decision to combine the prognostic models to produce a “categorical” estimate of survival (in terms of “days,” “weeks,” or “months+”) can also be considered to be a strength of the study. The resulting PiPS algorithms produce an estimate of survival that is clinically meaningful and that can be directly compared with clinicians’ own estimates.
Weaknesses of study
Our study also had several limitations. Although we tried to study all evaluable patients, some were not accessible because of “gatekeeping” by clinical staff. The phenomenon of “gatekeeping” is a common problem in palliative care studies,30 31 reflecting a desire by clinical staff to protect patients from the perceived burden of involvement in clinical studies. Such selective recruitment may distort findings if it renders the research sample unrepresentative of the population of interest. Aware that this was a potential limitation to our study, we invested considerable effort in accurately documenting “flows” of patients through the research process. Most previous palliative prognostic studies have not needed to rely on patients’ consent before recruitment or have reported only the results from “evaluable” patients.12 13 18 32 In contrast, we recorded data on all referrals to participating units and attempted to document the reasons why we could not evaluate some patients.
The fact that many potentially eligible patients for our study were non-evaluable was disappointing, but this reflects the reality of dealing with patients who have advanced disease. We did not find any significant differences in age, sex, or diagnoses between evaluable and non-evaluable patients. The most common reasons why patients were not evaluated were that they were under the care of the palliative care team for only a very short time (n=1899), so insufficient time was available to obtain informed consent; that the healthcare team deemed that discussion of prognosis would be too distressing (n=1102); or that the patient declined the offer from the healthcare team to talk to a researcher about the study (n=578). All of these reasons for failure to evaluate patients with the PiPS score also pertain to “real world” situations. We could therefore argue that the sample of patients recruited to our study is representative of the sort of patients who would be offered a PiPS evaluation in clinical practice.
Another potential limitation of our study is that the results have not yet been evaluated in an independent population. Developing prognostic models and testing them in the same population is known to produce over-optimism. The usual approach to cross validation is to develop the model in one half of the dataset (the training set) and to validate it in the other half (the testing set). Bootstrapping is an alternative cross validation technique that maximises the study data. It is a re-sampling procedure involving taking repeated samples (with replacement) from the dataset. The bootstrap validation in this study allowed us to “correct” for over-optimism in the development of the models and still to show that the PiPS predictions are robust. However, external validation in a separate cohort is still needed to confirm the predictive accuracy of the proposed models.
Our study was limited to determining the statistical accuracy of the PiPS prognostic algorithms. We did not assess whether introduction of the PiPS scores into clinical practice results in any demonstrable improvements in patient care. Future studies should include an assessment of the clinical utility as well as the statistical accuracy of the PiPS scores.
Relation to other studies
The PiPS models offer some definite advantages over existing methods of predicting survival in this population of patients. The most common method of prognostication is a clinician’s prediction of survival. However, as this is a subjective assessment, predictions cannot be standardised across sites or over time. This lack of consistency means that relying on clinicians’ predictions of survival is not an entirely suitable method of identifying patients for inclusion in clinical trials, for referral to specialist services, or for discussions with patients and families about the future. The PiPS-A scoring algorithms, although no more accurate than a multi-professional prognostic estimate, do provide an objective and repeatable output and would at the very least provide an excellent reproducible method for assessing eligibility for research studies. When blood results are available, the PiPS-B scores are significantly more accurate than either a doctor’s or a nurse’s estimate and thus could provide a robust rationale for making many decisions about treatment.
Other than clinicians’ estimates, the most widely studied method of predicting survival in patients with advanced cancer is probably the palliative prognostic score.12 13 This score is calculated from six variables, one of which is itself a clinician’s prediction. Each variable is allotted a “partial score” that depends on the size of the regression coefficient. Total palliative prognostic scores can range between 0 and 17.5 and define three prognostic categories representing different probabilities of surviving 30 days (<30%, 30-70%, and >70%). The palliative prognostic score has been subsequently validated in several clinical studies.33 34 35 36 37 One of its drawbacks is the heavy weighting that is attributed to the clinician’s prediction of survival; the partial score for the clinician’s prediction can represent 8.8/17.5 (49%) of the total score. Moreover, clinicians are required to provide an unrealistically accurate prognostic estimate in two week intervals. A patient who is expected to live for three to four weeks thus scores 1.5 points more than a patient expected to survive for five to six weeks and 3.5 points more than a patient expected for live for seven to 10 weeks. Most clinicians are reluctant to offer prognostic estimates with this degree of precision. Another limitation of the palliative prognostic score is the difficulty of converting the prognostic categories into meaningful clinical information. How should a clinician (or patient) interpret the information that they have a 30-70% chance of surviving 30 days? PiPS overcomes these problems.
Meaning of study
The PiPS-A score can be calculated for any patient with advanced cancer who is no longer receiving disease modifying treatment, and it is at least as good as, but not significantly better than, a clinician’s estimate of survival. The PiPS-B prognostic estimate requires a blood test, but it is statistically significantly better than the best uni-professional estimate of survival. Although the absolute increase in prognostic accuracy is not large, PiPS-B shares with PiPS-A the advantages of being independent of the clinician’s opinion and of being reproducible and comparable across settings. In a clinical context, we believe that PiPS estimates would usually be used to inform and augment clinicians’ own subjective estimates (rather than to replace them). However, our study suggests that the PiPS-B estimate of survival could now act as the “benchmark” against which new prognostic tools are assessed. We hope that this will act as the first step in an incremental process of improving prognostic accuracy as novel indicators are tested against, and then incorporated into, the PiPS models.
Unanswered questions and further research
Although the PiPS prognostic scores have been developed in a robust fashion, further validation work is needed before the scales can be recommended for use in routine clinical practice. The clinical information needed to calculate the PiPS scores is easily obtained. However, although the scores can be calculated with minimal arithmetical effort, they cannot be directly or easily calculated at the bedside. We have therefore produced a simple computer based interface for the PiPS instruments.38 We plan to develop this tool so that the algorithms will be available across a range of platforms and will allow users to rapidly convert clinical data into prognostic estimate categories.
As previously described, we have tested the PiPS models by using cross validation techniques, but their accuracy has not yet been assessed in independent datasets. The validation of prognostic tools emerged as the highest priority among clinicians and researchers in a recent consensus workshop on prognostication.39 We are planning a large prospective multicentre external validation study to include an assessment of users’ (clinicians, patients, and carers) views about the models and the best way for data to be presented to accurately reflect the degree of uncertainty inherent in the models.
What is already known on this topic
Prognostic information is valued by patients, carers, and healthcare professionals
Clinicians’ predictions of survival are the mainstay of current practice but are unreliable, over-optimistic, and subjective
What this study adds
Two prognostic scores have been created, both of which are able to predict whether patients will survive for “days,” “weeks,” or “months”
Both scores are independent of clinicians’ subjective estimates of survival, and both are at least as accurate as a clinician’s estimate
One of the prognostic scores (which requires a blood test) is significantly better than an individual doctor’s or nurse’s prediction, but neither scale is significantly more accurate than a multi-professional estimate of survival
Cite this as: BMJ 2011;343:d4920
We thank the following colleagues for their help with this study: Rehana Bakawala, Mike Bennett, Teresa Beynon, Cath Blinman, Patricia Brayden, Helen Brunskill, Kate Crossland, Alison Cubbitt, Rachel Glascott, Anita Griggs, Anne Harbison, Debra Hart, Philip Lomax, Caroline Lucas, Wendy Makin, Oliver Minton, Paul Perkins, Marek Plaskota, Dai Roberts, Katie Richies, Susan Salt, Ileana Samanidis, Margaret Saunders, Jennifer Todd, Catherine Waight, Nicola Wilderspin, Gail Wiley, and Julie Young. We also thank John Ellershaw for chairing the steering committee and Robert Godsill for providing a service user’s perspective. Thanks go to Rosie Head for administrative support and data management. Thanks also go to the following hospices and palliative care units for their participation in the study: Arthur Rank House (Cambridge), Worcestershire Royal Hospital, St John’s Hospice (Lancaster), Gloucestershire Hospitals NHS Foundation Trust, Pasque Hospice (Luton), Guy’s and St Thomas’ NHS Foundation Trust (London), Princess Alice Hospice (Esher), Bolton Hospice, St Catherine’s Hospice (Crawley), St George’s Hospital NHS Trust (London), Surrey and Sussex Healthcare NHS Trust, St Ann’s Hospice (Manchester), Christie Hospital NHS Foundation Trust (Manchester), Nightingale Macmillan Unit (Derby), Trinity Hospice (London), and Trinity Hospice (Blackpool).
Contributions: PCS, BG, VK, CT, CR, LK, and SB contributed to the conception and design of the study. CR, MG, and BG contributed to the analysis of data. All authors contributed to the interpretation of data, the drafting or revising of the manuscript, and final approval for publication. PCS is the guarantor.
Funding: This study was funded by Cancer Research UK (grant number C11075/A6126). SB is funded by Macmillan Cancer Support and the NIHR CLAHRC (Collaborations for Leadership in Applied Health Research and Care) for Cambridgeshire and Peterborough.
Competing interests: All authors have completed the Unified Competing Interest form at www.icmje.org/coi_disclosure.pdf (available on request from the corresponding author) and declare: support from Cancer Research UK (CRUK) for the submitted work; no financial relationships with any organisations that might have an interest in the submitted work in the previous three years; no other relationships or activities that could appear to have influenced the submitted work.
Ethical approval: This study was approved by the Wandsworth Multi-centre Research Ethics Committee. Site specific approval and research and development approval were obtained for participating units. Exemption from Section 60 of the Health and Social Care Act (2001) was obtained from the Patient Information Advisory Group (PIAG). This allows the records of incompetent patients to be accessed without explicit consent so that study variables can be recorded and patients “flagged” for mortality purposes with the NHS Information Centre. Consent was given by competent patients and assent by the relatives/carers of incompetent patients.
Data sharing: No additional data available.
This is an open-access article distributed under the terms of the Creative Commons Attribution Non-commercial License, which permits use, distribution, and reproduction in any medium, provided the original work is properly cited, the use is non commercial and is otherwise in compliance with the license. See: http://creativecommons.org/licenses/by-nc/2.0/ and http://creativecommons.org/licenses/by-nc/2.0/legalcode.