Journal of Epidemiology and Preventive Medicine

Monitoring Quality of Care in Men Diagnosed with Prostate Cancer: Developing Consensus Quality Indicators Using Modified-Delphi Methodology

Download PDF

Published Date: October 05, 2015

Monitoring Quality of Care in Men Diagnosed with Prostate Cancer: Developing Consensus Quality Indicators Using Modified-Delphi Methodology

Sue M Evans1*Denise Lin1, Dragan Ilic1, Jeremy Millar1,2, Declan Murphy3 and Joanne Dean1

1Department of Epidemiology and Preventive Medicine, Monash University, Melbourne Victoria, Australia

2William Buckland Radiotherapy Centre, Alfred Hospital, Melbourne, Victoria, Australia

3Australian Prostate Cancer Research Centre, Epworth Healthcare, Melbourne, Victoria, Australia

*Corresponding author: Sue M Evans, Monash University, Department of Epidemiology and Preventive Medicine, Tel: (+61) 399-030-017, (+61) 408-510-921; E-mail: sue.evans@monash.edu

Citation: Evans SM, Lin D, Ilic D, Millar J, Murphy D, et al. (2015) Monitoring Quality of Care in Men Diagnosed with Prostate Cancer: Developing Consensus Quality Indicators Using Modified-Delphi Methodology. J Epid Prev Med 1(2): 109. Doi: http://dx.doi.org/10.19104/jepm.2015.109 

       

Abstract

 

Objective: To develop a core set of clinical indicators to measure the quality of care provided to men with prostate cancer.

Design: A modified Delphi study involving interviews of key informants and two rounds of survey to obtain consensus on the indicator set.

Setting: Melbourne, Australia

Participants: n=20, including specialists involved in prostate cancer management (urology, radiation oncology, medical oncology, nursing, psychology, palliative care) epidemiologists, scientists, consumers and a policy advisor.

Intervention(s): A literature review was undertaken to identify potential quality indicators. Interviews were undertaken to ensure completeness of the set and explore potential for inclusion of novel indicators. Survey of Delphi panel participants was conducted to refine the list.

Main Outcome Measure(s): Items with panel agreement ≥ 60% for reliability and capacity to be objectively assessed and with a median validity score of ≥ 8 (scale ranged from 1 (not important) -9 (very important)).

Results: Of the total 104 proposed indicators, the panel retained 4/20 structural indicators, 15/46 process measures and 7/37 outcome indicators.

Conclusions: Indicators that scored highly in validity, reliability and objectivity included documentation of clinical stage, PSA level at diagnosis, surgical outcomes (rates of death, wound infection/bacteraemia and positive surgical margin), traditional measures of quality-of-care (10- and 15- year clinical and/or biochemical disease-free survival) and patient assessed post-treatment function using a validated survey instrument.

Keywords: Prostatic neoplasm; Quality indicator; Delphi; Consensus; Guidelines

Top ↑

List Of Abbreviations

 

PCR: Prostate Cancer Registry; PSA: Prostate Specific Antigen; EBRT: External Beam Radiation Therapy; ADT: Androgen Deprivation Therapy

Top ↑

Background

 

Prostate cancer is the most commonly diagnosed non-skin cancer among Australian males and its incidence and prevalence is growing [1]. In 2012 an estimated 18,560 new cases were diagnosed in Australia [2]. According to a study investigating trends in prostate cancer incidence across 40 countries, Australia and New Zealand had the highest age-standardised incident rate of 104 new cases per 100,000 population in 2008 [3]. More recent statistics suggest that in 2009 this rate was as high as 172 cases per 100,000 men [1]. Prostate cancer carries a huge economic burden to society. It is the most costly cancer, with a burden more than twice that of breast cancer, and three more than lung cancer [4].

Unlike many other cancers, many patients live with prostate cancer for decades and die from unrelated causes. Australian data indicates that 92% of men diagnosed with prostate cancer are alive five years later [5]. Recently published Victorian data demonstrate that the vast majority of men (93%) are diagnosed with localised disease and three quarters have low- or intermediate-risk disease [6]. Of those with low- or intermediate-risk disease, three-quarters will seek some form of active treatment in the initial 12 months post diagnosis. Treatment may consist of active surveillance, surgery (prostatectomy), radiotherapy, brachytherapy, hormone deprivation or a combination of these. To date, there is no conclusive evidence that any one treatment holds a survival advantage over another [7].

Given the high prevalence of prostate cancer in the community, its significant economic burden, the many treatment options available and the lack of evidence showing a survival advantage among treatment modalities, increasing attention is being paid to assessing whether quality-of-care and quality-of-life outcomes vary according to treatment and providers.

Best practice guidelines have been developed to guide prostate cancer management and clinical indicators have been devised to measure how well prostate cancer care aligns with these guidelines [8-11]. Researchers in the US have led quality indicator development using a multi-step process involving literature review, focus groups with patients and family members, interviews with clinical experts and panel discussion from the fields of urology, radiation oncology, medical oncology and health services research [12]. Proposed indicators were developed against criteria such as whether the indicator (1) had strong scientific basis; (2) had good face validity; (3) could be collected from a medical record, cancer registry, validated instrument or systematically recorded data source; and (4) was likely to be reliably reported. Failure to document information about the indicator would be a marker of poor quality [12]. As a result of this work, a consensus list of 49 quality indicators were developed from which a subset of 30 were tested in 770 private and public health care settings in the US providing external beam radiotherapy and surgery [13]. This identified considerable deficits in care, particularly for men undergoing surgery.

In Australia, this US-led work on quality indicator development provided valuable information to assist in the development of quality indicators for a population-based Prostate Cancer Registry (PCR). The PCR was established as a clinical quality registry to assess the impact of patient, clinician- and health service-related factors on morbidity and mortality following a diagnosis of prostate cancer [14]. The registry was developed following recognition through the state-based cancer registry that there was variation in survival following a diagnosis of prostate cancer according to regional location of diagnosis, and that reasons for this were not apparent [15].

Following a literature review and with expert input by a panel of three epidemiologists, a radiation oncologist, medical oncologist, two surgeons and a clinical pharmacologist, a set of indicators was developed: one measuring structure of the health service, six measuring processes of care and five assessing health outcomes were reviewed, as well as those developed by specialty groups. In 2012, the PCR Steering Committee determined that the initial quality indicator set ought to be more formally evaluated by a wider panel of stakeholders including clinicians, patients and scientists.

The aim of this paper is to outline the process used to develop quality indicators to measure quality and safety of prostate cancer care and detail the quality indicators selected for collection by the PCR.

Top ↑

Methods

 

Literature Review

A list of proposed variables and covariates was collated by examining the available literature. Guidelines from the United Kingdom [16], United States [17], Australia [11] and Europe [8] were reviewed.

Interviews

Purposive sampling was used to identify experts in the field of urology, medical oncology, radiation oncology, pathology, psychology and nursing and a consumer to interview to discuss existing indicators, explore potential for novel indicators and provide insight into current and future directions for management of prostate cancer. Following the literature review, semi-structured interviews were conducted using a topic guide in person and by telephone. In these interviews informants were asked to review and provide comment on the proposed draft set of indicators the panel would vote on; advise whether the Delphi panel information accompanying the survey was adequate to enable panel members to know what was expected of them; and advise whether there were areas of practice or indicators which ought to be included but which were not adequately covered with an appropriate indicator. All interviewed participants were invited to contribute to the Delphi process.

Delphi Process

Following the interviews, a panel of people with knowledge of various aspects of prostate cancer disease were invited to participate as members of a Delphi panel to anonymously assess whether the proposed indicator had good face and validity and whether data to construct the indicator could be reliably and objectively be collected from the medical record. Panel members were provided with definitions to assist them in their determination. Validity was defined as the extent to which the indicator captures an aspect of quality that is widely regarded as important and subject to health system control; reliability referred to the extent to which there was confidence in data collectors being able to consistently reproduce the data; and objectivity was defined as the extent to which data could be collected without influence/opinion of others. The invited panel members all contributed as members of an expert working group convened to develop a national prostate cancer registry. In addition, an international leader in prostate cancer research was invited and two additional urologists who participated in the interviews. Panel members were not informed of the identity of any other members of the panel. The Delphi method was originally developed at the RAND Corporation to systematically solicit the view of experts related to national defence [18]. The Delphi technique is well recognised as a tool for solving problems in health care settings [19]. It utilises a panel of experts to gain consensus of opinion through a series of questionnaires, interspersed with feedback [20].

Two rounds of online questionnaires were conducted. Members were asked to provide de-identified responses to multi-choice questions and free-text comments developed from the reviewed guidelines, literature search and semi-structured interviews. The responses were analysed by the research group and comments from the first round were incorporated in to the second round.

Each member was asked to rate each indicator on a scale of 1-9 to evaluate whether meeting the indicator would reflect high quality-of-care or, conversely, that not meeting the indicator would reflect poor quality-of-care. The indicators’ reliability and objectivity were also assessed with a yes/no/I don't know designation. A proposed indicator was considered valid if consensus was achieved, defined as a median score of 8 or above—with no disagreement according to the IPRAS (Interpercentile Range Adjusted for Symmetry), calculated with the formula provided in the RAND Appropriateness Method User's Manual [9]. An indicator with a median score of 8 or above was considered reliable and objective if the proportion of panel members who responded “yes” was more than 60%. An indicator with a median score of 8 or above, but with fewer than 60% of panel members considering it reliable or objective, was rejected.

In the free-text comment boxes, the experts were asked to provide suggestions for additional indicators and to present short justifications for a 1 or 9 rating.

Only positive outcomes and suggestions for new indicators were included in the second round, which used the same criteria for inclusion as the first round. Everyone who was sent an invitation in round one was also invited again to contribute in round two. Results were analysed by Microsoft Excel 2007 (Redmond, WA, Microsoft Corporation 2007) to calculate the median scores and IPRAS. This project was approved by the Monash University Human research Ethics Committee (LR CF12/1848 – 2012001023).

Consent

Written informed consent was provided by all participants involved in the project. A copy of the written consent is available for review by the Editor of this journal.

Top ↑

Results

 

Participation

Twenty-three people were invited to participate in the Delphi panel; of whom 20 accepted the invitation. Table 1 provides details of the specialities involved and the years of relevant experience for each group. One participant did not provide details of their experience. All participated in round one but three did not contribute to round two (85% response rate).

Table 1: Specialists involved in the Delphi Panel and years’ experience in the field 

 

Results from first round

The first round presented the panel with the initial list of 104 proposed quality indicators from the literature review and expert interviews. Using a framework for classifying quality of care created by Donabedian [21], these comprised of 20 structure, 47 process and 37 outcome measures and 8 covariates (Table 2 presents a summary with detailed description in appendix A). Structural indicators measure attributes of the settings in which care occurs, process indicators measure what is done in giving and receiving care and outcomes denote the effects of the care on the health status of patients and populations.

Table 2: Proposed indicators included in Round 1 of the Delphi process

 

Indicators with a median of seven or less were excluded for re-rating in the second round. At the end of round one, six new indicators were proposed and, in total 60 indicators (13 structure, 23 process and 24 outcome measures of quality) and five covariates qualified for re-rating in the second round. Table 3 provides a summary of indicators retained and suggested for entry into round two with details of indicators provided in appendix A.

Table 3: Indicators agreed upon after Round 1 of the Delphi process *Four indicators were newly proposed in Round 1and carried into Round 2. **Two indicators were newly proposed in Round 1 and carried into Round 2

 

Final endorsed indicators

Using the same criteria as in round one, at the end of round two, 26 indicators and five covariates were endorsed by the panel (4 structure, 15 process and 7 outcomes measures). The final lists of indicators are displayed in Table 4. The panel retained four of the 20 proposed structural indicators (20%), 15 out of the 46 proposed process measures (33%) and seven out of 37 proposed outcome indicators (19%).

Table 4: Quality indicators endorsed by the Delphi panel to monitor prostate cancer care. † predicts the ten-year mortality for a patient who may have a range of conditions such as heart disease, AIDS, or cancer (a total of 22 conditions). Each condition is assigned with a score of 1,2,3 or 6.

 

Five out of eight covariates were agreed upon to be important to control for when assessing quality of care, namely patient age and stage of disease at diagnosis (including PSA and Gleason score), family history of prostate cancer in first degree relatives, co-morbidities and use of adjuvant or neoadjuvant hormone treatment. The three exclusions were family history of prostate cancer in second degree relatives, personal history of other cancers and health insurance coverage.

Top ↑

Discussion

 

In this study, 31 indicators were endorsed as important measures of quality and safety when assessing prostate cancer management. In total there were four structural indicators, 19 process-of-care indicators and eight outcome indicators. Indicators that scored highly in validity, reliability and objectivity included documentation of clinical stage, PSA level at diagnosis, surgical outcomes (rates of death, wound infection/bacteraemia and positive surgical margin), traditional measures of quality-of-care (10- and 15- year clinical and/or biochemical disease-free survival) and patient assessed post-treatment function using a validated survey instrument. In general, our findings are very similar to those identified by a panel of experts in the United States and reported by Spencer et al [12] in 2003. However, we identified some indicators which are perhaps more pertinent in the Australian landscape, namely variation of access to services and consequently outcomes between regional and metropolitan patients, and also positive surgical margins, which the US panel did not endorse.

With regard to the structural indicators, the Delphi panel rejected patient-volume-of-prostatectomy, volume-of-external-beam-radiotherapy and of-seed-brachytherapy as quality indicators. This is at odds with a systematic review which suggested that provider volume (both hospital and surgeon) is an acceptable surrogate measure for quality-of-care in uro-oncological procedures [10]. In explaining why the panel did not enforce volume indictors, it may reflect the heterogeneity of the group or the contradicting evidence of a volume: quality relationship across all fields of medicine. Perhaps if a specialist urological panel were convened these indicators might have been endorsed. The structural indicator receiving greatest support was “Having access to MDM [multidisciplinary team meeting] decisions and outcomes”. Notwithstanding this expert support, a national audit of implementation of multidisciplinary cancer care in Australia suggested that this feature is not being applied or documented consistently [22]. Two-thirds of the 155 hospitals surveyed did not have a multidisciplinary team, and of those with one, one-quarter did not document the recommendations in the patient record.

Many panel members commented that hospital accreditation should not be included as a structural quality indicator, as the association between accreditation and quality-of-care was tenuous at best, in their view. A systematic review conducted in 2011 found insufficient evidence that external inspection to review compliance with Standards had any impact on improved patient outcomes and improved healthcare organisational and professional behaviour [23]. On the other hand, access to specialist pathologists for prostate specimens was prominently endorsed with panel members citing many occasions where initial pathology reports made by non-prostate-specialist pathologists disagreed with the reviewed opinion from specialist pathologists. The rationale for including it as an indicator was that treatment decisions were largely determined on the basis of histopathological reports. This is supported in the literature with a number of studies demonstrating discordant opinion on biopsy results, especially for extra prostatic extension and surgical margins, when specialist uro-pathologists’ findings were compared with those of general pathologists [24-26]. Differences in pathological findings have been shown to translate into different treatment provided with prognostic impact. While accessibility of uro-oncology nurses and interpreters have been endorsed as important indicators, the quality of such services has been highlighted as being much more difficult to assess objectively. Similarly, while the value of MDMs was recognised, some panel members question the thoroughness of MDM discussions, particularly if they are poorly attended, or if there are large numbers of patients to be discussed.

With regard to the “process indicators”, documentation of each of TNM staging, PSA and Gleason score were strongly endorsed, as expected. Other measures which had strong support assessed the quality of patient-centred care, with a large proportion of the final list focused on communication and full disclosure with the patient as well as patient-assessed baseline and post-treatment function. Equity indicators also feature prominently, with access to expertise and variations in outcomes between metropolitan versus regional patients and public versus private patients considered important when assessing quality and consistency-of-care. Studies have shown that patient-centred care is important in decision preparation, satisfaction and regret, by using appropriate language and formats for communication, fully preparing patients for tests and treatments and meeting the patients' needs for involvement in decision-making [27-29]. A review of patient decision making for localised prostate cancer found that the physician’s recommendation plays a significant role in influencing the patient’s choice of management and that most men will select the first treatment recommended to them. It may be that that other modalities are not discussed fully [30].

Of the eight retained outcome indicators, three are confined to patients undergoing radical prostatectomy (indicators 24, 25 and 26). There was a high level of support for documentation/assessment of longer term survival and disease recurrence, in keeping with increasing survival and hence increasing likelihood of recurrence in light of new technologies and treatment regimens. Assessment of post-treatment urinary, bowel and sexual function with a validated survey instrument was strongly supported, but patient satisfaction with those functions and treatment choice were not. Patient satisfaction was cited as too variable depending on how and when they are asked and their variable interpretation on what constituted success in the view of our panel. Furthermore, panel members raised the concern that these measures are influenced by other competing health risks and could act as a confounder and would not be directly attributable to the treatment toxicities.

Well-documented limitations of a Delphi process apply to our study: Delphi panels are not random and may not be representative of the expert groups included; there is no evidence of reliability of the results; and the existence of consensus does not mean that the correct answer has been found. Limitations of our particular study include the very lengthy round one questionnaire and comments being conveyed by proxy through round two instead of in person. One panellist did not complete round one of the survey due to the questionnaire length. Round one was necessarily long, as to include as many potential indicators as possible. Three panel members who contributed in round one did not contribute in round two. Future researchers could consider focus groups to eliminate certain measures before the questionnaire process if a long list is collated after the literature review.

Participants were not exposed to all controversial issues or general comments after round one, although all were informed indirectly through the round two surveys by the addition of new indicators and modifications of existing indicators. Had focus groups been conducted to discuss clinical indicators which were the subject of contention in more depth, some indicators ruled out may have been retained. A literature review to identify all quality indicators was performed using only the MEDLINE database. Further exploration on other databases, and a more specified search strategy may have identified further novel indicators that were not included in the initial process. Finally, since the indicators were developed new guidelines have been released and contemporary recommendations were not included in the proposed indicator set. These included recommendations relating to pre-biopsy and T and N staging imaging, management of an initial negative biopsy, active surveillance, the most effective radical prostatectomy method, combination of methods including EBRT and brachytherapy, and ADT and EBRT, ADT delivery and management of ADT side effects [31]. These updated recommendations highlight the need for quality indicators to be reviewed and updated on a regular basis to ensure they remain contemporary and relevant.

Top ↑

Conclusion

 

Through this project we have identified that the indicators originally selected for inclusion in the PCR remain relevant and important when assessed by a wider stakeholder group. The addition of 20 new indicators will provide a challenge to collect, and for this reason further work is required to quantify both the economic and data collection burden associated with these proposed indicators. However, if collection of the indicators translates to meaningful improvement in both quality of care and quality of life for men with prostate cancer, then this burden should be surmountable.

Top ↑

Acknowledgements

 

Dr Lin, A/Professor Evans and Ms Dean are supported through a Priority-driven Collaborative Research Scheme application funded by Cancer Australia (APP 1010384). Authors would like to acknowledge the contribution of panel members who participated in the Delphi panel.

Top ↑

References

 

  1. Australian Institute of Health and Welfare. Cancer in Australia: an overview 2012. Canberra, ACT: Australian Government; 2012.
  2. Australian Institute of Health and Welfare. cancer: key facts. Cancer 2013 [cited 2013 11 August ]; Available from: http://www.aihw.gov.au/cancer
  3. Center MM, Jemal A, Lortet-Tieulent J, Ward E, Ferlay J, Brawley O, et al. International variation in prostate cancer incidence and mortality rates. Eur Urol. 2012;61(6):1079-92. doi: 10.1016/j.eururo.2012.02.054.
  4. Glass P, Tracey E, Smetanin P, Kobak P, Pavilichev A, Bis J. Lives at Risk for Cancer in New South Wales 2007-2036: a health economics study of cancer in New South Wales. Eveleigh, NSW: Cancer Institute NSW; December 2008
  5. Australian Institute of Health and Welfare. Cancer Survival and Prevalence in Australia from 1982 to 2010. Canberra; 2012
  6. Evans SM, Millar JL, Davis ID, Murphy DG, Bolton DM, Giles GG, et al. Patterns of care for men diagnosed with prostate cancer in Victoria from 2008 to 2011. Med J Aust. 2013: 198(10):540-5
  7. Wilt TJ, MacDonald R, Rutks I, Shamliyan TA, Taylor BC, Kane RL. Systematic review: comparative effectiveness and harms of treatments for clinically localized prostate cancer. Ann Intern Med. 2008;148(6):435-48.
  8. Heidenreich A, Bellmunt J, Bolla M, Joniau S, Mason M, Matveev V, et al. EAU guidelines on prostate cancer. Part 1: screening, diagnosis, and treatment of clinically localised disease. Eur Urol. 2011;59(1):61-71. doi: 10.1016/j.eururo.2010.10.039.
  9. Fitch K, Bernstein SJ, Aguilar MD, Burnand B, Lacalle JR, Lazaro P, et al. The RAND/UCLA Appropriateness Method User's Manual. Santa Monica: The RAND Corporation; 2001.
  10. Nuttall M, van der Meulen J, Phillips N, Sharpin C, Gillatt D, McIntosh G, et al. A systematic review and critique of the literature relating hospital or surgeon volume to health outcomes for 3 urological cancer procedures. J Urol. 2004;172(6 Pt 1):2145-52.
  11. Australian Cancer Network Working Party on Management of Localised Prostate Cancer. Canberra: Commonwealth of Australia; 2002.
  12. Spencer BA, Steinberg M, Malin J, Adams J, Litwin MS. Quality-of-care indicators for early-stage prostate cancer. J Clin Oncol. 2003;21(10):1928-36.
  13. Miller DC, Spencer BA, Ritchey J, Stewart AK, Dunn RL, Sandler HM, et al. Treatment choice and quality of care for men with localized prostate cancer. Med Care. 2007;45(5):401-9.
  14. Evans SM, Millar JL, Wood JM, Davis ID, Bolton D, Giles GG, et al. The Prostate Cancer Registry: monitoring patterns and quality of care for men diagnosed with prostate cancer. BJU Int. 2013;111(4 Pt B):E158-66. doi: 10.1111/j.1464-410X.2012.11530.x.
  15. Giles GT, V (eds). Canstat: Prostate Cancer Anti-Cancer Council of Victoria; 2000 March 2000. Report No.: No 30
  16. National Collaborating Centre for Cancer. London: National Institute for Clinical Excellence (NICE); 2008 February. Report No.: NICE clinical guideline 58
  17. Mohler JL, Armstrong AJ, Bahnson RR, Boston B, Busby JE, D'Amico AV, et al. Prostate cancer, Version 3.2012: featured updates to the NCCN guidelines. J Natl Compr Canc Netw. 2012;10(9):1081-7.
  18. Dalkey NC. The Delphi Method: an Experimental Study of Group Opinion. Santa Monica, CA: RAND; 1969
  19. Fink A, Kosecoff J, Chassin M, Brook RH. Consensus Methods: Characteristics and Guidelines for Use. Am J Public Health. 1984: 74(9):979-83
  20. Linstone H, Turoff Me. The Delphi Method: Techniques and Applications, Reading, MA: Addison-Wesley, 1975
  21. Donabedian A. The quality of care. How can it be assessed? JAMA. 1988;260(12):1743-8.
  22. Wilcoxon H, Luxford K, Saunders C, Peterson J, Zorbas H, National Breast and Ovarian Cancer Centre's Multidisciplinary Care Audit Steering Committee. Multidisciplinary cancer care in Australia: a national audit highlights gaps in care and medico-legal risk for clinicians. Asia Pac J Clin Oncol. 2011;7(1):34-40. doi: 10.1111/j.1743-7563.2010.01369.x.
  23. Flodgren G, Pomey M-P, Taber SA, Eccles MP. Effectiveness of external inspection of compliance with standards in improving healthcare organisation behaviour, healthcare professional behavior or patient outcomes. Cochrane Database Syst Rev. 2011;(11):CD008992. doi: 10.1002/14651858.CD008992.pub2.
  24. Evans AJ, Henry PC, Van der Kwast TH, Tkachuk DC, Watson K, Lockwood GA, et al. Interobserver variability between expert urologic pathologists for extraprostatic extension and surgical margin status in radical prostatectomy specimens. Am J Surg Pathol. 2008;32(10):1503-12. doi: 10.1097/PAS.0b013e31817fb3a0.
  25. Goodman M, Ward KC, Osunkoya AO, Datta MW, Luthringer D, Young AN, et al. Frequency and determinants of disagreement and error in gleason scores: a population-based study of prostate cancer. Prostate. 2012;72(13):1389-98. doi: 10.1002/pros.22484.
  26. van der Kwast TH, Collette L, Van Poppel H, Van Cangh P, Vekemans K, DaPozzo L, et al. Impact of pathology review of stage and margin status of radical prostatectomy specimens (EORTC trial 22911). Virchows Arch. 2006;449(4):428-34.
  27. Berry DL, Wang Q, Halpenny B, Hong F. Decision preparation, satisfaction and regret in a multi-center sample of men with newly diagnosed localized prostate cancer. Patient Educ Couns. 2012;88(2):262-7. doi: 10.1016/j.pec.2012.04.002.
  28. Sinfield P, Baker R, Agarwal S, Tarrant C. Patient-centred care: What are the experiences of prostate cancer patients and their partners? Patient Educ Couns. 2008;73(1):91-6. doi: 10.1016/j.pec.2008.05.001.
  29. Song L, Chen RC, Bensen JT, Knafl GJ, Nielsen ME, Farnan L, et al. Who makes the decision regarding the treatment of clinically localized prostate cancer--the patient or physician?: results from a population-based study. Cancer. 2013;119(2):421-8. doi: 10.1002/cncr.27738.
  30. Zeliadt SB, Ramsey SD, Penson DF, Hall IJ, Ekwueme DU, Stroud L, et al. Why do men choose one treatment over another?: a review of patient decision making for localized prostate cancer. Cancer. 2006;106(9):1865-74.
  31. National Institute for Health and Care Excellence. Prostate cancer: diagnosis and treatment. 2014 [cited 2015 4 September 2015]; NICE clinical guideline 175.Available from: http://www.nice.org.uk/guidance/cg175.

Top ↑

Copyright: © 2015 Evans SM, et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.