Email updates

Keep up to date with the latest news and content from BMC Health Services Research and BioMed Central.

Open Access Research article

Evaluating maternity care using national administrative health datasets: How are statistics affected by the quality of data on method of delivery?

Hannah E Knight12*, Ipek Gurol-Urganci12, Tahir A Mahmood1, Allan Templeton1, David Richmond1, Jan H van der Meulen12 and David A Cromwell12

Author Affiliations

1 Office for Research and Clinical Audit, Lindsay Stewart R&D Centre, Royal College of Obstetricians and Gynaecologists, London 4RG, NW1, UK

2 Department of Health Services Research and Policy, London School of Hygiene and Tropical Medicine, London, WC1H 9SH, UK

For all author emails, please log on.

BMC Health Services Research 2013, 13:200  doi:10.1186/1472-6963-13-200

The electronic version of this article is the complete one and can be found online at: http://www.biomedcentral.com/1472-6963/13/200


Received:12 December 2012
Accepted:27 May 2013
Published:30 May 2013

© 2013 Knight et al.; licensee BioMed Central Ltd.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Background

Information on maternity services is increasingly derived from national administrative health data. We evaluated how statistics on maternity care in England were affected by the completeness and consistency of data on “method of delivery” in a national dataset.

Methods

Singleton deliveries occurring between April 2009 and March 2010 in English NHS trusts were extracted from the Hospital Episode Statistics (HES) database. In HES, method of delivery can be entered twice: 1) as a procedure code in core fields, and 2) in supplementary maternity fields. We examined overall consistency of these data sources at a national level and among individual trusts. The impact of different analysis rules for handling inconsistent data was then examined using three maternity statistics: emergency caesarean section (CS) rate; third/fourth degree tear rate amongst instrumental deliveries, and elective CS rate for breech presentation.

Results

We identified 629,049 singleton deliveries. Method of delivery was not entered as a procedure or in the supplementary fields in 0.8% and 12.5% of records, respectively. In 545,594 records containing both data items, method of delivery was coded consistently in 96.3% (kappa = 0.93; p < 0.001). Eleven of 136 NHS trusts had comparatively poor consistency (<92%) suggesting systematic data entry errors. The different analysis rules had a small effect on the statistics at a national level but the effect could be substantial for individual NHS trusts. The elective CS rate for breech was most sensitive to the chosen analysis rule.

Conclusions

Organisational maternity statistics are sensitive to inconsistencies in data on method of delivery, and publications of quality indicators should describe how such data were handled. Overall, method of delivery is coded consistently in English administrative health data.

Keywords:
Administrative health data; Maternity statistics; Method of delivery; Procedure codes; HES

Background

Countries which have administrative health data collection systems are increasingly using this information to produce maternity statistics at both local and national levels [1-3]. In the US, the Agency for Healthcare Research and Quality (AHRQ) developed a set of quality indicators based on administrative health data which included several areas of obstetric care [2]. These indicators have supported both national and local quality initiatives, and have been piloted in other developed countries including the UK, Canada, Spain, and Australia [4]. However, data quality remains a key concern for users of administrative maternity data and validation exercises are required to determine its accuracy and reliability prior to analysis [5].

In England, maternity statistics are produced by a number of organisations using the Hospital Episode Statistics (HES) database [6-9]. HES contains records on all patients admitted to English NHS hospitals, with data being extracted from local patient administration systems. The core fields of a HES record hold data on patient demographics and can capture up to 20 diagnoses and 24 procedures per episode of care. Delivery records can also capture supplementary data on the pregnancy and delivery, such as length of gestation, onset of labour, method of delivery and birth weight, for up to 9 babies. Not all delivery records contain this supplementary information (the ‘maternity tail’), although the percentage of records with a complete maternity tail has improved over time.

A number of quality indicators for hospital maternity services require method of delivery for their construction, for example the caesarean section rate (where it is required for the numerator), and the rate of third/fourth degree perineal tears amongst women delivering vaginally (where it is used in the denominator). Despite the importance of data on method of delivery, there is no current information on the quality of this data in HES. This is of concern because there are two ways in which method of delivery can be recorded in HES, and it is not clear which is the preferred data source. Until 2006, the UK Department of Health published figures on the consistency of the two sources of method of delivery for each hospital [10]. In addition, the Department of Health used to conduct extensive cleaning of HES data before its release for secondary analysis, but this has been replaced by a simpler data validation process.

This paper describes an evaluation of how statistics on maternity care in English hospitals are affected by the completeness and consistency of data on method of delivery. The completeness and internal consistency of HES method of delivery data were evaluated at a national level and by NHS trust (hospital organisation). We then assessed how different analysis rules for handling poor quality HES data influenced a selection of maternity statistics.

Methods

We extracted from the HES database records of women who delivered in English NHS acute trusts between 1 April 2009 and 31 March 2010. Records were defined as relating to a delivery if “method of delivery” information was found in any procedure field and/or the maternity tail field. Table  1 maps the Office of Population Census and Surveys (OPCS) procedure codes R17-25 on to the maternity tail codes used to define method of delivery [11]. The definitions for each category are equivalent. However, the OPCS codes are entered by clinical coders based on the discharge notes, whereas the maternity tail data is typically populated directly from the electronic maternity information system, which is completed by midwives.

Table 1. Correspondence between OPCS procedure delivery codes and maternity taildelmethdelivery codes

The analysis was limited to singleton deliveries. Records were excluded if they contained an International Classification of Diseases (ICD-10) diagnosis code for a multiple delivery (O30.1, Z37.2-.7 or Z38.3-.8) in any diagnosis field or the record contained data on more than one baby in the maternity tail.

Method of delivery was defined using a seven-category classification (Table  1). Both OPCS and maternity tail coding systems define ‘elective caesareans’ as prelabour caesarean sections and ‘emergency caesareans’ as an intrapartum caesarean sections. On inspection, a small number of hospitals had a value of “9” (other) in the maternity tail field for all deliveries, or seemed to have used this code to indicate an ‘unknown’ method of delivery. Consequently, if an NHS trust had values of “9” in the maternity tail field for more than 5% of their delivery episodes, all these values was re-coded as missing.

Data analysis

To examine data completeness for each NHS trust, we calculated the proportion of women for whom the method of delivery was recorded in a) the procedure fields, and b) the maternity tail. This analysis included all singleton delivery records. The subsequent analysis of coding consistency was restricted to women whose records contained information on method of delivery in both sources.

The mean rate of coding consistency was calculated by dividing the number of records with a consistent mode of delivery recorded in both the procedure field and the maternity tail by the total number of records containing valid information in both fields. We measured the overall level of coding agreement at a national level using the unweighted kappa (k) statistic. This measure ranges from 0 (a level of agreement no greater than would be obtained by chance) to 1 (perfect agreement). Values of k above 0.80 are generally considered to indicate excellent agreement [12].

We used funnel plots to examine variation among NHS trusts in the consistency of method of delivery coding [13,14]. The inner and outer control limits set at two and three standard deviations above and below the national average, respectively. The limits also took into account a measure of over-dispersion. This was derived using the random-effects method and incorporated 10% winsorisation to prevent the limits being widened excessively by extreme outliers [14]. The 0-5th percentiles were winsorised to the 5th percentile and the 95-100th percentiles were winsorised to the 95th percentiles.

We selected three maternity statistics to investigate the impact of using different analysis rules for handling inconsistent data. These were selected to represent the various categories of maternity statistic that require method of delivery:

•Emergency caesarean section rate (where method of delivery is the numerator);

•Third and fourth degree perineal tear rate amongst instrumental deliveries (where method of delivery is the denominator), and

•Elective caesarean section rate for breech presentation (where method of delivery affects both numerator and denominator).

Breech presentation was defined using ICD-10 codes O32.1, O64.1, O80.1 and O83.0-1 and/or OPCS or maternity tail codes for breech vaginal deliveries. We defined third and fourth degree tears as records with an ICD-10 code for third or fourth degree perineal laceration (O70.1; O70.2) and/or an OPCS procedure code for their repair (R32.1; R32.2).

Five versions of each statistic were produced using different analysis rules for dealing with inconsistencies in the method of delivery data (see Table  2 for definitions). To investigate the impact of these different rules on trust-level maternity statistics, we used mean-difference plots [15] to assess the agreement between two sets of figures, namely, figures derived using data on method of delivery from only the procedure fields (the approach currently used by the NHS Information Centre) [10] and figures derived using only those records for which the procedure and maternity tail data were in agreement (the most restrictive of the five analysis rules). STATA 11 (TX: StataCorp LP) was used for all statistical calculations.

Table 2. Impact of mode of delivery definition on resulting maternity statistics: three case studies

Results

Completeness of method of delivery codes

We identified 629,049 singleton deliveries in 151 English NHS trusts between 1 April 2009 and 31 March 2010. Among these, 545,594 records (86.7%) had method of delivery entered in both the procedure and maternity tail fields (Figure  1).

thumbnailFigure 1. Flow chart.

Method of delivery was mostly commonly entered as a procedure code, being omitted in just 4,850 records (0.8%) overall. All but four NHS trusts had a “method of delivery” procedure code in more than 95% of their deliveries, and in no trust was this code available in less than 90% of deliveries. In contrast, 78,605 records (12.5%) had method of delivery missing from the maternity tail. Only 96 of the 151 NHS trusts had a maternity tail “delivery” code in more than 95% of their deliveries, and seven NHS trusts had no information on delivery method in the maternity tail of their records.

Overall coding consistency

Among the 545,594 singleton deliveries with information in both the procedure and maternity tail fields, method of delivery was coded consistently in 96.3% records (kappa = 0.93, p < 0.001) using the seven category coding framework (Table  3). The overall rate of each delivery method differed by between 0 and 0.5% (e.g. the overall emergency caesarean section rate was 13.9% (76,004/545,594) from procedure codes and 13.7% (74,539/545,594) from maternity tail codes). However, coding inconsistencies had a large relative effect on the proportion of breech vaginal deliveries because it was an uncommon method of delivery. There were nearly twice as many records having this method of delivery in the maternity tail (3,170) compared to the procedures field (1,809) (Table  3).

Table 3. Consistency of Method of delivery in English NHS trusts in 2009/10 as defined using OPCS delivery code and the maternity tail delmeth code

Among all coding disagreements, 39% were inconsistencies between elective and emergency caesarean section (=[4,131 + 3,890]/20,402), while 19% were inconsistencies between instrumental and non-instrumental vaginal delivery (= [1,573 + 1,481 + 414 + 493]/20,402). A further 9% of inconsistencies were related to the type of instrument used to assist the delivery of the baby (=[1,573 + 1,481 + 414 + 493]/20,402) (see Table  3).

Variation in coding consistency between NHS hospital trusts

Figure  2 shows the variation in coding consistency among the 136 NHS trusts that had more than 500 delivery records containing both procedure and maternity tail codes. Eleven NHS trusts had levels of coding consistency lower than 92%, which was poorer performance than would be expected from random variation alone. The coding inconsistencies in these trusts appeared to be occurring systematically, accounting for 31% of all emergency/elective caesarean section discrepancies, 42% of all forceps/vacuum delivery discrepancies, 28% of all instrumental/non-instrumental delivery discrepancies, and 99% of all breech/vacuum delivery discrepancies.

thumbnailFigure 2. Funnel plot showing consistency between OPCS mode of delivery code and delmeth code for English NHS trusts (2009/10). The English average was calculated by dividing the number of records with consistent mode of delivery recorded in both fields by the total number of records containing information about mode of delivery in both fields.

The 11 NHS trusts with “poor” data quality accounted for 38,100 (7%) of the 545,594 singleton deliveries. Removing these trusts from the analysis improved the overall level of coding agreement from 96.3% (kappa = 0.93, p < 0.001) to 97.4% (kappa = 0.95, p < 0.001).

Impact of rules for handling data inconsistencies on maternity statistics

Table  2 shows the impact of using different analysis rules upon the three selected maternity statistics. At a national level, the different definitions had the smallest impact on the overall rate of third/fourth degree perineal tears amongst instrumental deliveries, with only 0.05% difference between the lowest and highest estimates. For the emergency caesarean section rate, the difference was almost 1%.

The most unstable statistic was the elective caesarean section rate among all women with breech presentation, with the estimated proportion ranging between 46.4% and 52.6% depending on which analysis rule was used. The inconsistencies in the definition of elective caesarean section affected the numerator, while the denominator was affected by the poor consistency in the definition of breech delivery.

Figure  3 shows the difference between the figures derived using two analysis rules at the level of individual NHS trusts. For the majority of NHS trusts, there was little difference between the emergency caesarean section rate and the third/fourth degree perineal tear rate amongst instrumental deliveries. The standard deviation (SD) of the difference was 1.6% and 1.0%, respectively. The spread of differences was larger for the elective caesarean section rate for breech presentation, with the SD of the differences being 5.5%. This reflects its comparatively smaller sample size in relation to the other statistics. Further analysis of these results showed that, for most trusts, the differences arose from changes in hospital sample size due to incomplete maternity tail data rather than inconsistencies of coding (Figure  2). Nonetheless, for each statistic, the different analysis rules produced very different figures for some NHS trusts, and these were typically those with poorer levels of coding consistency.

thumbnailFigure 3. Mean difference plots: Impact of using different method of delivery definitions on trust-level rates of : a) emergency caesarean section; b) third and fourth degree perineal tears amongst instrumental deliveries, and c) elective caesarean section for breech presentation. Definition 1: Use all episodes with an OPCS method of delivery code & base method of delivery definitions on OPCS codes alone; Definition 3: Uses only episodes in which both OPCS and maternity tail codes are present and in agreement; base method of delivery definitions on agreed method of delivery code.

Discussion and conclusion

This study evaluated the completeness and internal consistency of data on method of delivery within the HES database and how the accuracy of this data could affect different maternity statistics. We found that the procedure fields contained the most complete information on method of delivery, being available in 99.2% of records. They were also more consistently complete across all NHS trusts. The completeness of maternity tail information was considerably lower, and was missing entirely for seven NHS trusts.

When information was available in both sources, there was a high level of agreement between the method of delivery codes overall. Inconsistent coding was a problem in a minority of NHS trusts, with only 11 out of 136 trusts showing divergent coding practices. It was, therefore, not surprising that, at a national level, different rules for handling inconsistent data had a small effect on the derived statistics. Nonetheless, the degree of sensitivity varied across the statistics tested.

The variation in the level of data completeness and coding consistencies across NHS trusts meant that, for all statistics tested, the differences in the estimates produced by the alternative analysis rules were substantial for some trusts. These results highlight the need for a careful assessment of data quality and for the transparent reporting of how incomplete and inconsistent data are handled when producing maternity statistics, particularly at an organisational level.

This study included all singleton deliveries occurring in English NHS maternity units, providing a very large sample size for analysis and thereby reducing the risk of selection bias. We identified 629,049 singleton deliveries during the study time period, which represents approximately 97% of all hospital deliveries registered in England during 2009/10 by the Office for National Statistics [16]. Previous research shows that women with severe morbidity and prolonged hospitalisation are more likely to have delivery information missing from their records [17]. Although the loss of these women from analyses of mode of delivery is unlikely to make a difference, it would become extremely important if the data are used to assess maternal or perinatal morbidity and mortality.

A limitation of this evaluation is that it only assessed internal consistency. We did not attempt to validate the HES dataset by comparing a sample of records against hospital medical records. We are not aware of any studies that have specifically validated “method of delivery” coding in HES against hospital records, but studies of similar administrative health databases in other countries have reported high levels of agreement (kappa > 0.98, where stated) [18-21].

The seven method of delivery categories used in this study represent only one possible classification. The grouping was dictated by the OPCS procedure and maternity tail codes. A weakness of this classification is the definition of caesarean section as either elective or emergency. The 2004 NICE guideline recommended that the urgency of a caesarean section be indicated using the Lucas/National Confidential Enquiry into Patient Outcome and Death (NCEPOD) classification and noted that replacing the terms ‘emergency’ and ‘elective’ with its four grades of urgency would aid communication between health professionals [22]. Currently, the HES database is unable to capture this classification system.

Data quality is a concern for healthcare providers, managers and policy makers [23]. In England, the Care Quality Commission now mandates an annual audit of data quality within NHS trusts, [24] and a recent systematic review of coding accuracy in all types of routinely collected hospital discharge data found that coding accuracy rates have been improving [25]. Since 2002, the coding of primary diagnosis within HES has improved in accuracy from 73.8 per to 96.0% when compared against case notes [24].

The results of this study add to this work by addressing concerns about the quality of HES maternity data [26]. The high level of consistency in the recording of method of delivery overall supports its use for the construction of national maternity statistics. Coding disagreements were most common for the categories of emergency and elective caesarean section. Nonetheless, overall consistency was excellent between both emergency (kappa = 0.92; p < 0.001) and elective (kappa = 0.90; p < 0.001) caesarean section procedure and maternity tail codes. This supports a previous conclusion that coding errors were unlikely to account for the large variation in the rates of emergency caesarean section observed between NHS trusts [27].

At an NHS trust level, levels of consistency were high for the majority of organisations, which provides evidence to support the use of HES-based quality indicators for the purpose of comparing the performance of NHS trusts. However, our results illustrate the importance of addressing data quality within NHS trusts with divergent coding practices. The risk of organisations being mistakenly identified as “outliers” on performance indicators due to data errors is well-known. Our results suggest this risk is also increased by the sensitivity of maternity statistics to the analysis rules used to handle inconsistent data.

The study’s results also suggest that any publishers of maternity statistics should describe details of how data quality was assessed and incomplete and consistent data were handled in the analysis. In England, the Health and Social Care Information Centre (HSCIC) publishes maternity statistics at Strategic Health Authority, NHS trust and individual unit level annually [3]. This public body is England's central source of health and social care information and the value of its publications on maternity services would be enhanced if they again provided information on the level of agreement between data in the procedure fields and in the maternity tail.

Providing methodological information may be more problematic for commercial companies that supply hospitals with comparative measures of organisational performance given the need to balance transparency with the protection of intellectual property. Nonetheless, companies that provide maternity benchmarking services could be required to meet minimum standards of transparency as part of the conditions of access to administrative health data. Whilst national trends and local over time can be reported as long as the definitions used by these organisations remain the same, the definitions used are still important for interpretation.

Implications

Approaches to validate the use of administrative health data for maternity statistics commonly fall into two categories. They either check the consistency of the administrative health data against medical records [17-20,28] or against another source of maternity data such as national birth registers [29-31]. Such external validation studies can be time consuming, costly and technically challenging, as well as raising ethical and information governance issues related to access and data linkage. We used a particular feature of HES to examine its internal consistency and this is an example of how relationships within administrative health data can be used to identify organisations with divergent coding practices [32]. Whilst external validation should remain the “gold standard”, this approach to data quality assessment is simple to perform and has the potential to be developed more widely as a complementary technique.

Abbreviations

HES: Hospital Episode Statistics; ARHQ: Agency for Healthcare Research and Quality; IC: Information Centre; NHS: National Health Service; OPCS: Office of Population Census and Surveys; ICD-10: International Classification of Diseases, 10th Edition.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

DC, JvdM, TM, DR and AT conceived the idea. HK, DC, IG-U and JvdM designed the methodology. HK, IG-U and DC conducted the statistical analysis. HK wrote the manuscript. DC, JvdM, IG-U, TM, DR and AT commented on subsequent drafts and approved the final version. All authors read and approved the final manuscript.

Acknowledgements

We thank the Department of Health for providing the patient-level Hospital Episode Statistics data used in this study. Permission to use this data was granted by the NHS Information Centre. National maternity statistics and provider-level data are available via the HESonline website: http://www.hscic.gov.uk/hes.

References

  1. Roberts CL, Cameron CA, Bell JC, Algert CS, Morris JM: Measuring maternal morbidity in routinely collected health data: development and validation of a maternal morbidity outcome indicator.

    Med Care 2008, 46:786-794. PubMed Abstract | Publisher Full Text OpenURL

  2. Agency for Healthcare Research and Quality. 2012.

    http://www.qualityindicators.ahrq.gov/ webcite

    OpenURL

  3. Hospital Episode Statistics. 2013.

    http://www.hscic.gov.uk/hes webcite

    OpenURL

  4. Raleigh VS, Cooper J, Bremner SA, Scobie S: Patient safety indicators for England from hospital administrative data: case–control analysis and comparison with US data.

    BMJ 2008, 17(337):a1702. OpenURL

  5. Lain SJ, Hadfield RM, Raynes-Greenow CH, Ford JB, Mealing NM, Algert CS, Roberts CL: Quality of data in perinatal population health databases: a systematic review.

    Med Care 2012, 50(4):e7-e20. PubMed Abstract | Publisher Full Text OpenURL

  6. Dr Foster Intelligence. 2012.

    http://drfosterintelligence.co.uk/ webcite

    OpenURL

  7. CHKS. 2012.

    http://insight.chks.co.uk/index.php?id = 829 webcite

    OpenURL

  8. BirthChoiceUK. 2012.

    http://www.birthchoiceuk.com webcite

    OpenURL

  9. RCOG: Patterns of Maternity Care in English NHS Hospitals. London: RCOG; 2013.

    http://www.rcog.org.uk/files/rcog-corp/Patterns%20of%20Maternity%20Care%20in%20English%20NHS%20Hospitals%202011-12_0.pdf webcite

    OpenURL

  10. Health and Social Care Information Centre. NHS Maternity Statistics 2005–6; 2013.

    http://www.hscic.gov.uk/pubs/maternity0506 webcite

    OpenURL

  11. Health and Social Care Information Centre. NHS Maternity Statistics 2010–11 Explanatory notes; 2013:pp22-pp26.

    Available at: http://www.hscic.gov.uk/pubs/maternity1011 webcite

    OpenURL

  12. Petrie A, Sabin C: Medical Statistics at a Glance, Volume 39. 3rd edition. Oxford: Wiley-Blackwell Publishing; 2009:118. OpenURL

  13. Spiegelhalter DJ: Funnel plots for institutional comparison.

    Qual Saf Health Care 2002, 11:390-391. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  14. Spiegelhalter DJ: Funnel plots for comparing institutional performance.

    Stat Med 2005, 24:1185-1202. PubMed Abstract | Publisher Full Text OpenURL

  15. Bland JM, Altman DG: Statistical methods for assessing agreement between two methods of clinical measurement.

    Lancet 1986, 327:307-310. Publisher Full Text OpenURL

  16. Office for National Statistics:

    Births in England and Wales by parents' country of birth, 2010: Table 7. 2012.

    http://www.ons.gov.uk/ons/rel/vsob1/parents--country-of-birth--england-and-wales/2010/births-in-england-and-wales-by-parents--country-of-birth--2010.html webcite

    PubMed Abstract OpenURL

  17. Kuklina EV, Whiteman MK, Hillis SD, Jamieson DJ, Meikle SF, Posner SF, Marchbanks PA: An enhanced method for identifying obstetric deliveries: implications for estimating maternal morbidity.

    Matern Child Health J 2008, 12(4):469-477. PubMed Abstract | Publisher Full Text OpenURL

  18. Yasmeen S, Romano S, Schembri ME, Keyzer JM, Gilbert WM: Accuracy of obstetric diagnoses and procedures in hospital discharge data.

    Am J Obstet Gynecol 2006, 195:992-1001. OpenURL

  19. Roberts CL, Bell JC, Ford JB, Morris JM: Monitoring the quality of maternity care: how well are labour and delivery events reported in population health data?

    Paediatr Perinat Epidemiol 2009, 23:144-152. PubMed Abstract | Publisher Full Text OpenURL

  20. Lydon-Rochelle MT, Holt VL, Cárdenas V, Nelson JC, Easterling TR, Gardella C, Callaghan WM: The reporting of pre-existing maternal medical conditions and complications of pregnancy on birth certificates and in hospital discharge data.

    Am J Obstet Gynecol 2005, 193(1):125-134. PubMed Abstract | Publisher Full Text OpenURL

  21. Korst LM, Gregory KD, Gornbein JA: Elective primary caesarean delivery: accuracy of administrative data.

    Paediatr Perinat Epidemiol 2004, 18:112-119. PubMed Abstract | Publisher Full Text OpenURL

  22. National Collaborating Centre for Women’s and Children’s Health: Caesarean section. London: National Institute for Clinical Excellence; 2004. OpenURL

  23. Audit Commission: Data remember: improving the quality of patient-based information in the NHS 2002. London: Audit Commission; 2012.

    http://archive.audit-commission.gov.uk/auditcommission/sitecollectiondocuments/AuditCommissionReports/NationalStudies/dataremember.pdf webcite

    OpenURL

  24. Care Quality Commission

    2012.

    http://www.cqc.org.uk/organisations-we-regulate/registered-services/quality-and-risk-profiles-qrps webcite

  25. Burns EM, Rigby E, Mamidanna R, Bottle A, Aylin P, Ziprin P, Faiz D: Systematic review of discharge coding accuracy.

    J Public Health 2012, 34:138-148. Publisher Full Text OpenURL

  26. Brennan L, Watson M, Klaber R, Charles T: The importance of knowing context of hospital episode statistics when reconfiguring the NHS.

    BMJ 2012, 344:e2432. PubMed Abstract | Publisher Full Text OpenURL

  27. Bragg F, Cromwell DA, Edozien LC, Durol-Urganci I, Mahmood TA, Templeton A, van der Meulen JH: Variation in rates of caesarean section among English NHS trusts after accounting for maternal and clinical risk: cross sectional study.

    BMJ 2010, 341:c5065. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  28. Lain S, Roberts C, Hadfield R: How accurate is the reporting of obstetric haemorrhage in hospital discharge data? A validation study.

    Austral N Z J Obstet Gynecol 2008, 48:481-485. Publisher Full Text OpenURL

  29. Joseph KS, Fahey J: Validation of perinatal data in the Discharge Abstract Database of the Canadian Institute for Health Information.

    Chronic Dis Canada 2009, 29:96-100. OpenURL

  30. Dattani N, Datta-Nemdharry P, Macfarlane A: Linking maternity data for England, 2005–06: methods and data quality.

    Health Stat Q 2011, 49:53-79. PubMed Abstract | Publisher Full Text OpenURL

  31. Dattani N, Datta-Nemdharry P, Macfarlane A: Linking maternity data for England 2007: methods and data quality.

    Health Stat Q 2012, (53):4-21. OpenURL

  32. Johal A, Mitchell D, Lees T, Cromwell D, Van Der Meulen J: Use of Hospital Episode Statistics to investigate abdominal aortic aneurysm surgery.

    Br J Surg 2012, 99(1):66-72. PubMed Abstract | Publisher Full Text OpenURL

Pre-publication history

The pre-publication history for this paper can be accessed here:

http://www.biomedcentral.com/1472-6963/13/200/prepub