Statistics and Evidence-Based Healthcare





Introduction


Many doctors often equate statistics with the numbers and equations seen in research papers, but the term ‘statistics’ does not mean ‘numbers’; indeed, a competent statistical analysis of a paper should include non-numerical issues such as the nature of the sampling methods or the validity of a ‘gold standard’ diagnostic test. Furthermore, papers may be overflowing with numerical data but contain no statistics at all.


Statistics has been defined as the discipline concer-ned with:




  • Data collection and presentation



  • Inference from samples or experiments to the population at large



  • Modelling and analysis of complex systems



  • Broader issues to do with the application and interpretation of the above techniques in politics, management, the law, philosophy, ethics and the sciences.



One of the problems in getting to grips with statistics is that, in common with many other branches of medicine, it is becoming an increasingly sophisticated science, where standard textbooks appear to serve only those who are already members of its exclusive club. The analysis of large and complex datasets, and the techniques of mathematical modelling and statistical computer programming, are generally best left to the experts. However, the appraisal of many published articles relevant to the practicing obstetrician or gynaecologist can be greatly helped by an understanding of several basic principles, some of which are presented in this chapter.


Some Basic Statistical Principles


Sampling And Inference To The Population At Large


The most fundamental issue of statistics is that one is trying to relate data taken from a relatively small ‘sample’ to a much larger group where it would be impractical to collect all the available data. In medical statistics, this large and rather nebulous group of subjects is known as ‘the population’ and it can often be hard to define. This may seem obvious but understanding the concept of sampling is crucial in the interpretation of results from studies. The application of statistics to research is an attempt to ensure that the results from your sample are in general agreement with the results that would have occurred if you had been able to conduct the experiment on all relevant members of the population in question. Fig. 14.1 demonstrates this simple relationship, and it is intuitive to see that, as the sample size increases, it moves closer to representing the population. In general, as the size of the sample increases, the bias in the study decreases; however, this is not always the case, and in some unusual statistical settings the bias will remain, but these are beyond the scope of this chapter (see Peacock and Peacock, 2011).




Fig. 14.1


Representation of sample and population.


The most important ‘take-home’ message for you as an investigator is that, if you repeated your experiment, you would almost certainly get a different result. You may still draw the same conclusions from that result, but the numbers used in the statistical calculations would differ from sample to sample. On average, you would expect the results to be consistent, but when there is disagreement between sample and population, the following types of error (Type 1 and Type 2 errors) can occur and these should always be at the front of your mind when interpreting study results.


Type 1 Error


This occurs when ‘the sample’ used in your experiment generates a significant result for your hypothesis but there would not have been a significant result if you had performed the experiment on ‘the population’; in other words, it occurred by chance. When we set a ‘significant’ P value at .05, we are allowing a 5% chance of a type 1 error occurring for our study. To have zero chance of a type 1 error we would need to perform the experiment on ‘the population’ itself, and this is not possible as it implies recruiting an infinite number of subjects. The 5% level is entirely arbitrary and purely a convention that seems ‘reasonable’ in most research situations. Thus, P values should be interpreted cautiously, always bearing in mind the study design and the size of the difference observed between the comparative groups. P values are a continuum that runs from 0 to 1 and thresholds such as .05 are only used as a guide. On occasions, different levels are used, such as 1% ( P value of .01) and 10% ( P value of .10).


Another consideration when interpreting the results of a study is whether many comparisons have been made using the same sample. This is known as ‘multiple testing’ and it is a common flaw seen in a number of studies in the published literature (see Peacock and Peacock 2011 for further details on how to address this). By setting your P value for significance at .05, you are allowing a 5% chance of a type 1 error, that is, 1 in 20 statistical comparisons will produce a significant result by chance and thus, if a large number of variables in the dataset are tested for significance, there is a considerable risk of getting a type 1 error. Multiple testing often goes hand in hand with an unclear hypothesis and a poorly thought through study design. If you wish to test many outcomes and exposures using the same sample of patients, you need to account for this by setting yourself a more stringent P value for significance, for example, .01. In this case, you would need to make 100 comparisons in order for one of them to be significant by chance.


Type 2 Error


This occurs when ‘the sample’ used in your experiment fails to generate a significant result for your hypothesis but there would have been a significant result if you had performed the experiment on ‘the population’, that is, you have missed a real and possibly important effect. When we require a study to have 90% power, we are allowing a 10% chance that our sample will not detect a significant result that in truth exists in the population. This commonly occurs in small studies where there is insufficient power. Under-powered studies can be frustrating to interpret, particularly when there appears to be quite a large difference between the groups but the P value does not reach the magical threshold for acceptance as a significant result. Some statisticians argue that no under-powered studies should ever be undertaken as they cannot be interpreted and can raise ethical concerns in terms of unjustified patient research, and this would probably condense the world’s research output to a fraction of its current amount. Most research funders now demand power calculations, but sometimes the assumptions on which power calculations are based are grossly optimistic. However, in many cases a balance can still be achieved between attaining sufficient power and setting a pragmatic target for the sample size.


The Null And Active Hypotheses


It is always important to be able to define the null and active hypotheses for a study, and this means having clear definitions for both the outcome and the exposure or treatment. Under the null hypothesis, there is no difference between the groups that are being compared. This will tend to be the ‘default’ hypothesis unless the study sample accrues sufficient evidence to reject this null hypothesis and show that the active hypothesis is true.


Bias And Generalisability


When studies have been sampled in a non-random fashion, differences between the results from the sample and the true population can arise. Similarly, if treatments are allocated to patients non-randomly, estimates for differences between treatment groups can be biased. In other words, bias can arise if the sample is systematically unrepresentative of the population. However, bias and generalisability are not always the same thing. If one runs a well-powered randomised controlled trial (RCT), the results comparing randomised groups that are estimated from the sample are unlikely to be biased; however, they will only be generalisable to the sub-group of the population that met the inclusion criteria for the study. Thus, when interpreting the results from studies it is important to view them in relation to the study inclusion and exclusion criteria and the sampling methods that were employed when recruiting the subjects. There are many different types of bias and certain study designs are more prone to particular types of bias than others. This will be discussed in more depth in the section about types of study and experimental design ( pp 10 ).


Confidence Intervals, Accuracy And Precision


When interpreting a result from a sample, it is useful to express the result with a range of possible values that it might have taken if other samples of the same size had been selected. This range of values is called a confidence interval (CI), and we can set the level of confidence as a percentage; 95% confidence is typically used. For example, if we measure the birth weight of 50 babies and calculate a point estimate for the mean weight of 3360 g and a 95% CI of 3200 to 3520 g, this means that we can be 95% confident that, given this sample size, the true mean birth weight for all babies in the population relevant to this study lies somewhere between 3200 and 3520 g. In general, as the sample size increases, the CI becomes narrower. The term precision is used to describe how wide the CI is around the point estimate, whereas accuracy gives an indication of how close the point estimate from the sample is to the true unmeasurable population value and is therefore more related to bias or generalisability. The calculation of CIs will be discussed in more detail on p. 7 .


Primary and Secondary Outcomes


When designing a study, there will often be many research outcomes listed. In general, one or a few of these outcomes will be identified as the primary outcome(s). The primary outcome(s) is used to calculate the sample size for the overall study. A well-presented study will provide the P value along with the CI of the primary outcome(s) being measured, while the estimates of the secondary outcomes should be reported with CIs.


Independence And Matched Data


Many statistical tests make assumptions about the independence of the subjects analysed in the study. If data are not independent (e.g. the same mother can be included more than once in a study on childbirth) then account should be taken of this in the analysis. Many commonly used statistical tests assume that all observations are from separate individuals and, by including subjects more than once, you are making your sample less varied than it would be if subjects were only entered once. Similarly, if your study design selected cases and controls by matching them in terms of covariates such as age and ethnic group, then your analysis must account for this matching. In general, it is preferable not to match any subjects within a study as it is possible to adjust for potential differences between your groups at the analysis stage. Furthermore, in studies where subjects have been assessed before and after experiencing an exposure, they should be investigated in a ‘paired’ fashion by analysing the difference between the before and after measurements, as this accounts for the lack of independence between them. This also tends to improve the power of the study as within-patient differences tend to be less variable than absolute variation between patients.


Data Types, Distribution Assumptions and Parametric Tests


There are many different ways in which we can collate data on a subject of interest but, in general, data can be classified into the following types:




  • Quantitative or continuous – a continual spectrum of data measurements, for example, age, blood pressure, height or weight.



  • Ordinal – subjects are categorised into groups where there is some order to the categories, for example, mild, moderate or severe symptoms.



  • Categorical – subjects are categorised into groups but there is not necessarily any particular order to the categories, for example, ethnicity or country of birth.



  • Binary – this is a sub-group of ordinal and categorical data where there are just two possible categories, for example, pregnant or not pregnant, dead or alive.



  • Time-dependent data – where subjects have been followed up for different lengths of time, typically in cohort studies and RCTs when subjects have been recruited over an extended period of time. For example, the classification of a subject as pregnant or not pregnant may depend upon the length of their follow-up.



When analysing these data types, it is often necessary to make assumptions about how the data within our sample are likely to behave in relation to the population from which they came. In order to do this, it is helpful to assume a probability distribution which can be described by a mathematical equation, and this can then be used as a template to describe the sample data and make comparisons within it. There are many types of mathematical distribution used in statistics, but four of the most common ones are the binomial, Poisson, normal and chi-squared (χ 2 ) distributions:




  • The binomial distribution describes the probability distribution for binary data, and it relates to the common example of tossing a coin. For large sample sizes, the binomial distribution is very similar to the normal distribution and so the latter is often assumed in the statistical calculations.



  • The Poisson distribution can be assumed when investigating rates derived from time-to-event data, and it represents the idea that a certain event is occurring at a constant rate and thus, as we follow people through time, more events will occur. However, it should be noted that there are also more complex assumptions that are required when analysing time-to-event data, for example, Cox proportional hazards regression analysis is often used (see Peacock and Peacock, 2011).



  • The normal or Gaussian distribution is assumed when investigating measurements from continuous data, but it is also used as the basis for many aspects of medical statistics. More detail is provided for this distribution a little further on, as it is so important in understanding the application of statistics.



  • The chi-squared ( χ 2 ) distribution is derived by squaring the normal distribution, and it has particular properties that make it useful for investigating proportions from categorical, ordinal or binary data.



The Normal Distribution


The normal distribution is one of the most important and widely used probability distributions in medical statistics. It can be described by a rather complex mathematical equation; however, if it is plotted in terms of probability, we can see that it generates the famous ‘bell-shaped’ curve shown in Fig. 14.2 . The x-axis is standardised such that the mean corresponds to zero (the most probable value) with units of standard deviation (SD) falling above and below this value. It can be seen that 95% of the area under the curve lies between the points that fall 1.96 SDs on either side of the mean value, and this number is particularly important as we can use it to give us an indicator of the range of values that would incorporate 95% of all possible values. In some cases, you may wish to know the range of values that incorporate 90% or even 99% of all values, and these ranges correspond to the 1.65 and 2.58 SDs on either side of the mean, respectively.




Fig. 14.2


Probability distribution function for the normal distribution. The horizontal axis has been standardised such that zero corresponds to the mean with units of standard deviation above and below the mean.


The beauty of the normal distribution is that this symmetrical property around the mean value holds whether we are plotting the actual data points from our sample or whether we are plotting the results of the study if we had repeated it over and over again. In this scenario, we would end up with the ‘mean of the mean of the samples’, and the range of mean values can be represented by the sampling distribution. The term standard error is essentially the SD of this sampling distribution and it is used throughout statistics to calculate CIs around point estimates. An example of how to calculate a CI for the mean is given on p. 7 .


Parametric And Non-Parametric Tests


Parametric statistical tests are ones where assumptions are made about which mathematical distribution best represents the sample and the population from which it was taken. Non-parametric statistical tests are ones where no assumption has been made about the distribution of the data. See Table 14.1 ‘A guide to unifactorial statistical methods’. In general, parametric tests tend to be more powerful and sensitive than non-parametric tests and therefore tend to be preferred, as fewer observations are required to provide evidence in favour of the hypothesis if it is true. A typical example of a parametric test is the use of a Student’s t -test to compare the mean values of a continuous variable between two groups. One of the test assumptions is that the continuous data measured in the sample can be assumed to follow the normal distribution. If this assumption is not valid, then the non-parametric Mann–Whitney U test can be used, which ranks the observations in order of size and compares the proportions that fall above and below the median value for each of the groups in question. Thus, the Mann–Whitney U test is less sensitive to large outlying values but also less informative as observations above or below the median are all treated in the same way.



Table 14.1

A Guide to Unifactorial Statistical Methods

Reproduced from Peacock JL, Peacock PJ 2011 Oxford handbook of medical statistics, Oxford University Press, Oxford.
















































































Design or Aim of Study Type of Data/Assumptions Statistical Method
Compare Two Independent Samples
Compare two means Continuous, Normal distribution, same variance t test for two independent means
Compare two proportions Categorical, two categories, all expected values greater than 5 Chi-squared test
Compare two proportions Categorical, two categories, some expected values less than 5 Fisher’s exact test
Compare distributions Ordinal Wilcoxin two-sample signed rank test equivalent to Mann Whitney U test
Compare time to an event (e.g. survival) in two groups Continuous Logrank test
Compare Several Independent Samples
Compare several means Continuous, Normal distribution, same variance One-way analysis of variance
Compare time to an event (e.g. survival) in several groups Continuous Logrank test
Compare Differences in a Paired Sample
Test mean difference Continuous, Normal distribution for differences t test for two paired (matched) means
Compare two paired proportions Categorical, two categories (binary) McNemar’s test
Distribution of differences Ordinal, symmetrical distribution Wilcoxon matched pairs test
Distribution of differences Ordinal Sign test
Relationships Between Two Variables
Test strength of linear relations between two variables Continuous, at least one has Normal distribution Pearson’s correlation
Test strength of relationship between two variables Ordinal Spearman’s rank correlation, Kendall’s tau (if many ties)
Examine nature of linear relationship between two variables Continuous, residuals from Normal distribution, constant variance Simple linear regression
Test association between two categorical variables Categorical, more than two categories for either or both variables, at least 80% of expected frequencies greater than 5 Chi-squared test
Test for trend in proportions Categorical, one variable has two categories and the other has several categories which are ordered, sample greater than 30 Chi-squared test for trend


Deciding Whether to Use Parametric or Non-Parametric Tests


For binary data, the assumption of a binomial distribution may be valid, as according to Peacock and Peacock, 2011, the normal distribution ‘can be used as an approximation to the Binomial distribution when n is large. In precise this works if np and n(1−p) are both greater than 5 (where np and n(1−p) are number of successes and number of failures)’. When comparing proportions across binary, categorical or ordinal data, the chi-squared distribution is often assumed; however, if the numbers in the categories in a 2 × 2 table become very small (<5 in any one category) then it is often more appropriate to use Fisher’s exact test, which is described in any standard statistical textbook.


Probably the most common example of deciding whether to use a parametric or non-parametric test is when you want to know whether the continuous data in your sample can be assumed to follow a normal distribution. In general, for small samples of less than about 15 observations, it is not safe to assume the data are normally distributed, and non-parametric methods should generally be employed. However, it should be remembered that these tests are less powerful and the sample size is small, which will make the statistical results hard to interpret. If you have a reasonably large sample size, the first thing to do is to plot your data points on a scatter graph or group the data into bins and plot them on a histogram. Inspection of the graphs or histograms is the simplest way of assessing whether your distribution assumptions are valid. Deviations from the normal distribution can lead to significant skewness or kurtosis. Fig. 14.3 demonstrates histograms for data that follow a normal distribution or have a positively or negatively skewed distribution, and Fig. 14.4 shows how data can deviate from the classic ‘bell-shaped’ curve seen in the normal distribution and exhibit kurtosis. Kurtosis is concerned with the shape of the distribution and can have a considerable impact on the statistical analysis that you choose to perform on your data. When kurtosis is extreme, non-parametric tests should be used. It is worth noting that for data that are perfectly normally distributed, the mean, median and mode values are all the same, whereas for positively skewed data, the mean tends to be larger than the median and vice versa for negatively skewed data. When summarising skewed data (and also reporting descriptive results from a non-parametric test), it is often better to quote the median and interquartile range rather than the mean and SD, which are generally used for summarising normally distributed data. The way to calculate these summary statistics is described in the next section. In some cases, it helps to convert skewed data into another variable that can be assumed to follow the normal distribution (this is called transformation). For example, data that are positively skewed can often be manipulated into a more normally distributed format by transforming them onto the log scale; the t -test can then be used on the log-transformed data. There are more formal ways of testing your assumptions about the normal distribution, such as normal plots and Shapiro–Francia or Shapiro–Wilk tests, but these should be used cautiously and, if there is doubt, you should revert to using non-parametric methods.




Fig. 14.3


Histograms representing skewness of data. (A) Negatively skewed (median > mean). (B) Normally distributed (median = mean = mode). (C) Positively skewed (median < mean).



Fig. 14.4


Histograms representing kurtosis of data. (A) Lepto-kurtic (long-tailed). (B) Meso-kurtic (normally distributed). (C) Platy-kurtic (short-tailed).


Data Collection and Presentation


There are numerous ways in which data can be summarised and presented, and the choice depends on the type of data. The previous section defined the different types of data and the distributions that are often assumed to analyse them. Table 14.2 shows some typical methods for summarising and presenting different data types. All statistical software packages will perform the analysis for summary statistics, but the following simple example shows how the basic ones can be calculated for a set of data.



Table 14.2

Typical Summary Statistics and Presentation Methods for Describing Data




















Data Type Summary Statistics Presentation
Quantitative or continuous Mean or average, standard deviation, variance, standard error, confidence intervals, mode, median, range, interquartile range


  • Scatter plots



  • Line plots



  • Box and whisker plots



  • Histograms

Categorical, ordinal or binary


  • Mode



  • Percentage or risk



  • Odds




  • Histograms



  • Pie charts



  • Bar charts

Time-dependent event outcomes


  • Rate



  • Hazard (particular example for Cox modelling)




  • Life table or



  • Kaplan–Meier curves



For example, a study was conducted to investigate various aspects of pregnancy status, previous pregnancy, delivery type, birth weight and survival in a sample of 20 women recruited over 1 year and followed for 5 years.


Mean


This is the sum of all the observations divided by the number of observations.


For example, mean systolic blood pressure at 20 weeks’ gestation = (105 + 107 + 107 + … + 190 + 199)/20 = 144.35 mmHg.


Similarly, the mean age = 32.05 years.


Variance


This is an indication of the variability of the observations. Each observation is subtracted from the mean, squared, added up and divided by the number of observations, minus 1.


For example, variance of systolic blood pressure at 20 weeks’ gestation = [(105 − 144.35) 2 + (107 − 144.35) 2 + … + (190 − 144.35) 2 + (199 − 144.35) 2 ]/(20 − 1) = 804.66 mmHg.


Similarly, the variance of age = 71.94 years.


Standard Deviation


This is also an indication of the variability of the observations as it is the square root of the variance.


For example, SD of systolic blood pressure at 20 weeks’ gestation = <SPAN role=presentation tabIndex=0 id=MathJax-Element-1-Frame class=MathJax style="POSITION: relative" data-mathml='804.66′>804.66804.66
804 .66
= 28.37 mmHg.


Similarly, the SD of age = 8.48 years.


Standard Error Of The Mean


The standard error is used to indicate how well the sample mean measurement represents the true population mean value. Standard errors are used to calculate CIs (see next section).


For example, standard error of the mean (SEM) systolic blood pressure at 20 weeks’ gestation = SD divided by the square root of the number of observations = <SPAN role=presentation tabIndex=0 id=MathJax-Element-2-Frame class=MathJax style="POSITION: relative" data-mathml='28.37/20′>28.37/2028.37/20
28 .37 / 20
= 6.34 mmHg.


Similarly, the standard error for the mean age is 1.90 years.


Confidence Interval For The Mean


Earlier in the chapter, the characteristics of the normal distribution were discussed, and these properties are central to the construction of CIs. The most common CI is set at 95%, as this corresponds to a P value of .05. Once the standard error has been calculated, the 95% CI for the mean blood pressure can be constructed using the 1.96 multiplier described on page 4 . Thus, the point estimate with the 95% CI for the mean blood pressure is 144.35 ± (1.96 × 6.34) = 131.9 to 156.8 mmHg.


Similarly, the 95% CI around the mean age is 28.3 to 35.8 years.


If you wish to be more stringent with your data, you can set your P value threshold for statistical significance at .01, rather than .05, as this corresponds to a 99% CI. In this case, the 1.96 number increases to 2.58. Alternatively, a less stringent threshold would be a P value of .1 where the 1.96 value is decreased to 1.65 to generate a 90% CI.


Mode


This is the most common value in the dataset. It is typically used with categorical and ordinal data but can also be used for continuous data. The mode can become a more complicated parameter when the distribution of data has more than one most common value, for example, bi-modal, but this will not be discussed further here.


For example, the mode for ethnic group is white and the mode for delivery type is home delivery.


Median


The median is the midpoint of all the observations, indicating that 50% of the observations lie above and 50% lie below the median. Sometimes it is more appropriate to quote the median rather than the mean as it is less sensitive to large outlying values. Similarly, when data are skewed, it is generally better to quote the median rather than the mean.


For example, the median for blood pressure at 20 weeks’ gestation is midway between the 10th and 11th observation when arranged in rank order, that is, 146 mmHg.


Similarly, the median age is 32 years.


Range


This is the total range of values between the largest and the smallest observation. It indicates how widely varied the data are and is often quoted with the median.


For example, the range for blood pressure at 20 weeks’ gestation is 105 to 199 mmHg.


Similarly, the range for age is 20 to 49 years.


Interquartile Range


This is similar to the median, although the lower value indicates that 25% of the observations lie below it and the upper value indicates that 25% of the observations lie above it. Thus, it represents the central 50% range of values and is usually quoted with the median and often used to summarise skewed data.


For example, the The interquartile range for blood pressure at 20 weeks’ gestation is 118.5 to 163.5 mmHg.


Similarly, the interquartile range for age is 24.5 to 37 years.


Proportion And Risk


Table 14.3 is a 2 × 2 table that shows the results of delivery type by previous full-term pregnancy. The percentages can be used to represent the proportion of previous full-term pregnancies within hospital or home birth deliveries, that is, 3 of 7 hospital deliveries occurred in women who had a previous full-term pregnancy (42.9%) compared with 5 of 13 home deliveries (38.5%). Risks and percentages cannot be used for time-dependent data unless subjects have all been followed for the same length of time.



TABLE 14.3

Results of Hospital/Home Delivery and Previous Full-Term Pregnancy
























Previous Full-Term Pregnancy No Full-Term Pregnancy Total
Hospital 3 4 7
Home 5 8 13
Total 8 12 20

Only gold members can continue reading. Log In or Register to continue

Stay updated, free articles. Join our Telegram channel

Aug 6, 2023 | Posted by in OBSTETRICS | Comments Off on Statistics and Evidence-Based Healthcare

Full access? Get Clinical Tree

Get Clinical Tree app for offline access