ArticlePDF Available

Abstract and Figures

Biomedical research is seldom done with entire populations but rather with samples drawn from a population. Although we work with samples, our goal is to describe and draw inferences regarding the underlying population. It is possible to use a sample statistic and estimates of error in the sample to get a fair idea of the population parameter, not as a single value, but as a range of values. This range is the confidence interval (CI) which is estimated on the basis of a desired confidence level. Calculation of the CI of a sample statistic takes the general form: CI = Point estimate ± Margin of error, where the margin of error is given by the product of a critical value (z) derived from the standard normal curve and the standard error of point estimate. Calculation of the standard error varies depending on whether the sample statistic of interest is a mean, proportion, odds ratio (OR), and so on. The factors affecting the width of the CI include the desired confidence level, the sample size and the variability in the sample. Although the 95% CI is most often used in biomedical research, a CI can be calculated for any level of confidence. A 99% CI will be wider than 95% CI for the same sample. Conflict between clinical importance and statistical significance is an important issue in biomedical research. Clinical importance is best inferred by looking at the effect size, that is how much is the actual change or difference. However, statistical significance in terms of P only suggests whether there is any difference in probability terms. Use of the CI supplements the P value by providing an estimate of actual clinical effect. Of late, clinical trials are being designed specifically as superiority, non-inferiority or equivalence studies. The conclusions from these alternative trial designs are based on CI values rather than the P value from intergroup comparison.
Content may be subject to copyright.
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
Introduction
Biomedical research is seldom done with entire populations
but rather with samples drawn from a population. There
are various strategies for sampling, but, wherever feasible,
random sampling strategies are to be preferred since they
ensure that every member of the population has an equal
and fair chance of being selected for the study. Random
sampling also allows methods based on probability theory
to be applied to the data.
Although we work with samples, our goal is to
describe and draw inferences regarding the underlying
population. Values obtained from samples are referred to
as ‘sample statistics’ which we have to use to garner idea
of corresponding values in the underlying population, that
are referred to as ‘population parameters’. But how do we
do this? If we are doing ‘census’ type of studies, then the
measured values are directly the population parameters since
a census covers the entire population. However, if we are
studying samples, then what we have in our hand at study
end are the sample statistics. If we have a large enough and
adequately representative sample, it is logical to presume
that the sample statistics would be close to the ‘true values’,
that is the population parameters, but they would probably
not be identical to them. Strictly speaking, without doing
a census it is not possible to get true population values.
Practically speaking, it is possible to use a sample statistic
and estimates of error in the sample to get a fair idea of
Statistics Corner
Using the confidence interval confidently
Avijit Hazra
Department of Pharmacology, Institute of Postgraduate Medical Education & Research, Kolkata, India
Correspondence to: Dr. Avijit Hazra, MD. Department of Pharmacology, Institute of Postgraduate Medical Education & Research (IPGME&R), 244B
Acharya J. C. Bose Road, Kolkata 700020, India. Email: blowfans@yahoo.co.in.
Abstract: Biomedical research is seldom done with entire populations but rather with samples drawn from
a population. Although we work with samples, our goal is to describe and draw inferences regarding the
underlying population. It is possible to use a sample statistic and estimates of error in the sample to get a fair
idea of the population parameter, not as a single value, but as a range of values. This range is the condence
interval (CI) which is estimated on the basis of a desired condence level. Calculation of the CI of a sample
statistic takes the general form: CI = Point estimate ± Margin of error, where the margin of error is given
by the product of a critical value (z) derived from the standard normal curve and the standard error of point
estimate. Calculation of the standard error varies depending on whether the sample statistic of interest is a
mean, proportion, odds ratio (OR), and so on. The factors affecting the width of the CI include the desired
condence level, the sample size and the variability in the sample. Although the 95% CI is most often used
in biomedical research, a CI can be calculated for any level of condence. A 99% CI will be wider than 95%
CI for the same sample. Conict between clinical importance and statistical signicance is an important issue
in biomedical research. Clinical importance is best inferred by looking at the effect size, that is how much is
the actual change or difference. However, statistical signicance in terms of P only suggests whether there
is any difference in probability terms. Use of the CI supplements the P value by providing an estimate of
actual clinical effect. Of late, clinical trials are being designed specically as superiority, non-inferiority or
equivalence studies. The conclusions from these alternative trial designs are based on CI values rather than
the P value from intergroup comparison.
Keywords: Condence interval (CI); condence level; P value; statistical inference; clinical signicance
Submitted Aug 21, 2017. Accepted for publication Aug 29, 2017.
doi: 10.21037/jtd.2017.09.14
View this article at: http://dx.doi.org/10.21037/jtd.2017.09.14
4130
4126 Hazra. Using the CI confidently
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
the population parameter, not as a single value, but as a
range of values. This range is the condence interval (CI).
How well the sample statistic estimates the underlying
population value is always an issue. The CI addresses this
issue because it provides a range of values which is likely to
contain the population parameter of interest.
The CI is a descriptive statistics measure, but we can use it
to draw inferences regarding the underlying population (1).
In particular, they often offer a more dependable alternative
to conclusions based on the P value (2). They also indicate
the precision or reliability of our observations—the narrower
the CI of a sample statistic, the more reliable is our estimation
of the underlying population parameter. Wherever sampling
is involved, we can calculate CI. Thus we can calculate CI
of means, medians, proportions, odds ratios (ORs), relative
risks, numbers needed to treat, and so on. The concept of
the CI was introduced by Jerzy Neyman in a paper published
in 1937 (3). It has now gained wide acceptance although
many of us are not quite condent about it (4). To be fair it
is not an intuitive concept but requires some reection and
effort to understand, calculate and interpret correctly. In this
article we will look at these issues.
Meaning of CI
The CI of a statistic may be regarded as a range of values,
calculated from sample observations, that is likely to contain
the true population value with some degree of uncertainty.
Although the CI provides an estimate of the unknown
population parameter, the interval computed from a
particular sample does not necessarily include the true
value of the parameter. Therefore, CIs are constructed
at a confidence level, say 95%, selected by the user. This
implies that were the estimation process to be repeated over
and over with random samples from the same population,
then 95% of the calculated intervals would be expected
to contain the true value. Note that the stated condence
level is selected by the user and is not dependent on the
characteristics of the sample. Although the 95% CI is by far
the most commonly used, it is possible to calculate the CI
at any given level of condence, such as 90% or 99%. The
two ends of the CI are called limits or bounds.
CIs can be one or two-sided. A two-sided CI brackets the
population parameter from both below (lower bound) and
above (upper bound). A one-sided CI provides a boundary for
the population parameter either from above or below and thus
furnishes either an upper or a lower limit to its magnitude.
Calculation of CIs
Formulas for calculating CIs take the general form:
CI = Point estimate ± Margin of error
Point estimate ± Critical value (z) × Standard error of point
estimate
The point estimate refers to the statistic calculated from
sample data. The critical value or z value depends on the
condence level and is derived from the mathematics of the
standard normal curve. For condence levels of 90%, 95%
and 99% the z value is 1.65, 1.96 and 2.58, respectively.
The standard error depends on the sample size and the
dispersion in the variable of interest.
Calculation of the CI of the mean is relatively simple.
Here the formula is:
CI = Sample mean ± z value × Standard error of mean (SEM)
Sample mean ± z value × (Standard deviation/n)
If we are calculating the 95% CI of the mean, the
z value to be used would be 1.96. Table 1 provides a listing of
z values for various condence levels. The margin of error
depends on the size and variability of the sample. Naturally,
the error will be smaller if the sample size (n) is large or the
variability of the data [standard deviation (SD)] is less and
this is reected in the SEM.
Ideally the SD used in the calculation should be the
population SD. However, this is often unknown and if we are
dealing with a reasonably large (say n >100, or at least >30)
random sample, then the sample SD can be used as a fair
approximation of the population SD. If the sample is small
and one has to rely on the sample SD, then this requires
derivation of the CI using the t distribution rather than
the z value from the normal distribution. In this situation,
the z value is to be replaced with the appropriate critical
Table 1 Critical (z) values used in the calculation of confidence
intervals
Confidence level Critical (z) value to be used in confidence
interval calculation
50% 0.67449
75% 1.15035
90% 1.64485
95% 1.95996
97% 2.17009
99% 2.57583
99.9% 3.29053
4127Journal of Thoracic Disease, Vol 9, No 10 October 2017
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
value of the t distribution with (n–1) degrees of freedom.
Note that the Student’s t distribution, resembles the normal
distribution, although its precise shape depends on the
sample size. The required t value can be found from a t
distribution table included in most statistical textbooks. For
example, if the sample size is 25, the critical value for the
t distribution that corresponds to a 95% confidence level
with 24 degrees of freedom, is 2.064.
Let us now work out an example. The mean systolic
blood pressure (SBP) and diastolic blood pressure (DBP) of
72 randomly selected chest physicians aged over 50 is 134
and 88 mmHg, with SD of 5.2 and 4.5 mmHg. What is the
95% CIs for the blood pressure readings?
Here we will consider the sample SD as fair
approximation to the population SD. Therefore:
95%CI of SBP 134 1.96 (5.2/ 72) 134 1.20 . ., 132.8 to 135.2 mmHg
95%CI of DBP 88 1.96 (4.5 / 72) 88 1.04 . ., 86.9 to 89.0 mmHg
ie
ie
=±× =±
=±× =±
Thus the chest physicians appear to be non-hypertensive
at present although they still need to keep a regular check
on their blood pressure readings.
Unlike numerical variables, categorical variables are
summarized as counts or proportions, and we will now deal
with CI of a proportion. The formula for this is a bit more
intimidating, but is still manageable for manual calculation.
CI = Sample proportion (p) ± z value × Standard error of proportion
(1 - )p
p z value p n

±×


Let us work through an example. A radiology resident has
done a small observational study to nd out the sensitivity
and specicity of pleural effusion detected on digital chest
X-rays (CXR) in predicting the risk of malignancy among
subjects presenting with suggestive clinical findings. Her
data is summarized in Table 2.
The sensitivity of his diagnostic modality is [63/(63+25)]
×100, i.e., 71.59%, while specificity is [53/(33+53)]
×100, i.e., 61.63%; The 95% CI for the sensitivity would be:
0.7159 1.96 [0.7159(1 0.7159) /174] 0.7159 0.0670 . ., 64.89 to 78.29%ie±× = ±
And the 95% CI for the specicity would be:
0.6163 1.96 [0.6163(1 0.6163) /174] 0.6163 0.0723 . ., 54.40 to 68.86%ie±× = ±
In this same example, the odds of malignancy being
present when pleural effusion is detected on CXR is 63/33,
i.e., 1.9091, while the odds when pleural effusion is absent
is 25/53, i.e., 0.4717. Therefore the OR for detecting
malignancy when pleural effusion is present, compared to
when it is absent, would be 1.9091/0.4717, i.e., 4.047.
Calculation of the 95% CI of the OR requires a more
complicated formula, where we first derive the natural
logarithm (log to base e, or ln) of the sample OR and then
calculate its standard error. From this we derive the two
condence limits of the ln(OR), and then take their antilog
to derive the 95% CI of the OR.
The formula is:
ln(OR) 1.96 SE(ln(OR))
e±×
where,
Thus in the above example ln(OR) would be ln(4.047), i.e.,
1.3980. The SE of ln(OR) would be
, i.e., 0.3241. The 95% condence limits for
ln(OR) would be 1.3980±1.96×0.3241=1.3980±0.6352, i.e.,
0.7628 to 2.0332.
Taking antilog of these two boundaries, the 95% CI of
OR of 4.047 in this case would be 2.144 to 7.638. Note that
since this 95% CI of the OR does not span the value 1, it
implies that a pleural effusion detected by CXR in a patient
with clinical examination ndings suggestive of malignancy
is approximately 2.1 to 7.6 times as likely to indicate
malignancy than when it is not detected.
The reason why we resort to such an apparently complicated
formula is that ORs are not normally distributed. They
tend to be skewed towards the lower end of possible values.
As a result, we must take the natural log of the OR and rst
compute the condence limits on a logarithmic scale, and
Table 2 Radiology resident’s cross-tabulation of pleural effusion vis-a-vis chest malignancy data
Pleural effusion on chest X-ray
Final diagnosis of malignancy
Present Absent Row totals
Present 63 33 96
Absent 25 53 78
Column totals 88 86
4128 Hazra. Using the CI confidently
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
then convert them back to the normal OR scale. The same
applies to relative risks and other sample statistics that are
skewed in this manner.
Calculating CIs for distribution free statistics
From the examples above, it should be evident that CI take the
general form of Point estimate ± Margin of error, where the
margin of error is calculated by multiplying a critical value selected
as per the required condence limits with the standard error.
Standard errors cannot be calculated for distribution
free statistics. Nevertheless, CIs can be calculated and have
the same interpretation, that is they will present a range of
values with which the true population value is compatible.
In this case, the confidence limits are not necessarily
symmetric around the sample estimate and are given
by actual values in the sample that are chosen from the
applicable formula.
The formula for calculating the CI of the median is:
1.96 1.96
1
2 2 22
NN Nn
th ranked value to th ranked value− ++
For example, if there are 100 values in a sample data
set, the median will lie between 50th and 51st values when
arranged in ascending order. Applying the formula shown
above, the lower 95% confidence limit is indicated by
40.2 rank ordered value, while the upper 95% condence limit
is indicated by 60.8 rank ordered value. Since there are no actual
40.2 and 60.8 ranked values, we choose the ranks nearest to these
and values of these ranks then provide the approximate 95% CI
for the median. For the 100 value series, this will therefore be
the range indicated by the 40th to 61st rank ordered value.
For large samples, the CI for the median and other
quartiles can be determined on the basis of the binomial
distribution. You can see examples online (5).
Factors affecting the width of the CI
The width of the CI indicates the utility of our estimation
of the population parameter. Suppose the weather forecaster
uses the concept of the 99% CI to declare that tomorrows
maximum temperature is going to be anywhere between
1 to 50 . It is extremely unlikely that he or she will be
wrong but for you and me it would be utmost perplexing as
to what to wear outdoors tomorrow. On the other hand if he
says that the range is likely to be 20 to 30 on the basis of
the 95% CI, then he has more chance of being wrong but it
is easier for us to decide how we dress tomorrow. The factors
affecting the width of the CI include the desired condence
level, the sample size and the variability in the sample.
The width of the CI varies directly with the condence
level. A 99% CI would be wider than the corresponding
95% CI from the same sample. This stands to reason, since
a larger likelihood of containing the true population value
would lie with the wider interval.
A larger sample size expectedly will lead to a better estimate
of the population parameter and this is reected in a narrower
CI. The width of the CI is thus inversely related to the sample
size. In fact, required sample size calculation for some statistical
procedures is based on the acceptable width of the CI.
Variability in a random sample directly inuences the width
of the CI. A larger spread implies that it is more difficult to
reliably estimate population value without large amounts of data.
Thus as the variability in the data (often expressed as the SD)
increases, the CI also widens.
Practical use of the CI
In descriptive statistics, CIs reported along with point
estimates of the variables concerned, indicate the reliability
of the estimates. The 95% confidence level is often used,
though the 99% CI are used occasionally. At 99%, the width
of the CI will be larger but it is more likely to contain the true
population value, than the narrower 95% CI. Bioequivalence
testing makes use of the 90% CI. In such studies, we can
conclude that two formulations of the same drug are not
different from one another if the 90% CI of the ratios for
peak plasma concentration (Cmax) and area under the plasma
concentration time curve (AUC) of the two preparations
(test vs. reference) lies in the range 80–125%.
Conflict between clinical importance and statistical
significance is an important issue in biomedical research.
Clinical importance is best inferred by looking at the effect
size, that is how much is the actual change or difference.
However, statistical significance in terms of P only suggests
whether there is any difference in probability terms (6). One
way to combine statistical significance and effect sizes is to
report CIs. If a corresponding hypothesis test is performed, the
condence level is the complement of the level of signicance,
that is a 95% CI reects a signicance level of 0.05, while at
the same time providing an estimate of the ‘true’ value. Indeed,
if there is no overlap on comparing 95% CI surrounding point
estimates of the outcome variable in different groups, once
can conclude that statistically signicant difference exists. On
the other hand, even if there is small overlap, the difference
between groups may not be clinically signicant, irrespective
4129Journal of Thoracic Disease, Vol 9, No 10 October 2017
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
of the P value. Thus stating the CI shifts the interpretation
from a qualitative judgment about the role of chance to a
quantitative estimation of the biologic measure of effect (7).
Of late, clinical trials are being designed specifically as
superiority, non-inferiority or equivalence studies. The
conclusions from these alternative trial designs are based
on CI values rather than the P value from intergroup
comparison (8). CI around the outcome point estimate
for the test drug must fall wholly within a predefined
equivalence margin on both sides of the line of no
difference for establishing equivalence. For establishing
non-inferiority, the lower bound of the 95% CI for the test
drug must not cross the non-inferiority margin set a priori.
For establishing superiority, the lower bound of the 95% CI
for the test drug must lie beyond the line of no difference,
while the upper bound extends beyond the superiority
margin set a priori. Figure 1 summarizes these situations
graphically. Selection of these margins has to be done with
due care based on clinical judgment.
The concept of CIs and confidence levels are also used
in the calculation of sample size for prevalence surveys. The
margin of error selected by the surveyor determines the
acceptable deviation between the prevalence in the surveyed
section of the population and the prevalence in the entire
population. Thus, the margin of error implies a CI. The
condence level selected indicates how often the percentage of
the population that has the condition of interest is likely to lie
within the boundaries decided by the margin of error. Table 3
indicates how required sample size for population surveys
varies with acceptable margin of error and condence level.
Figure 1 The positioning of 95% confidence limits around the point estimate in the test intervention group to establish non-inferiority,
equivalence or superiority in clinical trials. The line of no difference is indicated by zero; negative delta indicates the non-inferiority margin or
the lower bound of the equivalence margin, while positive delta indicates upper bound of the equivalence margin or the superiority margin.
Control better 0 Test drug better
Non-inferiority Equivalence Superiority
Control better 0 Test drug better Control better 0 Test drug better
− ∆ − ∆ + ∆ + ∆
Table 3 Sample size required for surveys
Estimated
population size
Margin of error
Confidence level 95% Confidence level 99%
5% 2.5% 1% 5% 2.5% 1%
100 80 94 99 87 96 99
500 217 377 475 285 421 485
1,000 278 606 906 399 727 943
10,000 370 1332 4899 622 2098 6239
100,000 383 1513 8762 659 2585 14227
500,000 384 1532 9423 663 2640 16055
1,000,000 384 1534 9512 663 2647 16317
Sample size is larger for a lower margin of error or higher level of confidence. Once the estimated population size is very large (>100,000),
the sample size is not changing much.
4130 Hazra. Using the CI confidently
© Journal of Thoracic Disease. All rights reserved. J Thorac Dis 2017;9(10):4125-4130jtd.amegroups.com
Misconceptions regarding the CI
A 95% CI does not mean that 95% of the sample data lie
within that interval. A CI is not a range of plausible values
for the sample, rather it is an interval estimate of plausible
values for the population parameter.
It is natural to interpret a 95% CI as a range of values with
95% probability of containing the population parameter.
However, the proper interpretation is not that simple. The
true value of the population parameter is fixed, while the
width of the 95% CI based on a random sample will also vary
randomly. If we take repeated random samples of equal size
from the population, we will get a corresponding number of
95% CI values not all of which will contain the population
parameter—in fact only 95% of them can be expected to
contain the population parameter value. Thus the CI may not
always give an idea of the population parameter.
Selection of the acceptable condence level is arbitrary. We
often use the 95% CI in biological sciences, but this is a matter
of convention. A much higher level is often used in the physical
sciences. For instance the six sigma concept, the quality
improvement program that Motorola originated, and which
is now popular in many manufacturing companies, utilizes
a confidence level of 99.99966% (9). The engineers want to
eliminate all risk of manufacturing poor-quality products and
therefore work at this level of precision.
Finally, it is worthwhile to remember that the concept
of the CI was introduced to provide an answer to the
vexing issue in statistical inference of how to deal with
the uncertainty inherent in results derived from data that
represent randomly selected subset of a population. There
are other answers, notably that provided by Bayesian
inference in the form of credible intervals. Calculation of the
conventional CI depends on set rules that ensure that the
interval determined by the rule will include the true value of
the population parameter. This is the so called ‘frequentist’
approach. The Bayesian approach offers intervals that
can, subject to acceptance of interpretation of ‘probability’
as Bayesian probability, be interpreted as meaning that
the specific interval calculated from a given dataset has a
particular probability of including the true value, conditional
on the particular situation (10). Bayesian intervals treat their
bounds as fixed and the estimated parameter as a random
variable, whereas conventional approach treats confidence
limits as random variables and the population parameter as
a fixed value. Unlike the Bayesian method, the frequentist
method of computing CIs does not make use of any other
prior information regarding the location of the population
parameter. Thus, there is a philosophical difference between
the two approaches which we can address at a later stage.
Acknowledgements
None.
Footnote
Conicts of Interest: The author has no conicts of interest to
declare.
References
1. Altman DG. Why we need condence intervals. World J
Surg 2005;29:554-6.
2. Akobeng AK. Condence intervals and p-values in clinical
decision making. Acta Paediatr 2008;97:1004-7.
3. Neyman J. Outline of a theory of statistical estimation
based on the classical theory of probability. Philos Trans R
Soc Lond A 1937;236:333-80.
4. Hoekstra R, Morey RD, Rouder JN, et al. Robust
misinterpretation of condence intervals. Psychon Bull
Rev 2014;21:1157-64.
5. Bland M. Condence interval for a median and other
quartiles [Monograph on the internet]. Available online:
https://www-users.york.ac.uk/~mb55/intro/cicent.htm
6. Thiese MS, Ronna B, Ott U. P value interpretations and
considerations. J Thorac Dis 2016;8:E928-31.
7. Medina LS, Zurakowski D. Measurement variability and
condence intervals in medicine: why should radiologists
care? Radiology 2003;226:297-301.
8. Lesaffre E. Superiority, equivalence, and non-inferiority
trials. Bull NYU Hosp Jt Dis 2008;66:150-4.
9. Anonymous. What is a condence interval and why would
you want one? [Monograph on the internet]. Available online:
http://www.uxmatters.com/mt/archives/2011/11/what-is-a-
condence-interval-and-why-would-you-want-one.php
10. Jaynes ET. Condence intervals vs Bayesian intervals.
In: Harper WL, Hooker CA, editors. Foundations of
probability theory, statistical inference, and statistical
theories of science. Vol II. Dordrecht: D. Reidel
Publishing Company 1976:175-257.
Cite this article as: Hazra A. Using the confidence interval
condently. J Thorac Dis 2017;9(10):4125-4130. doi:10.21037/
jtd.2017.09.14
... The confidence interval (CI) provides a statistical range within which the true value of a population parameter, such as a mean or proportion, is estimated to lie with a specified level of confidence. A widely used and suitable value for CI is the 95% confidence interval, suggesting that if the same sampling method is repeated numerous times, about 95% of the intervals calculated would contain the true population parameter 57 . The mathematical expression for calculating a confidence interval is: ...
Article
Full-text available
The widespread availability of miniaturized wearable fitness trackers has enabled the monitoring of various essential health parameters. Utilizing wearable technology for precise emotion recognition during human and computer interactions can facilitate authentic, emotionally aware contextual communication. In this paper, an emotion recognition system is proposed for the first time to conduct an experimental analysis of both discrete and dimensional models. An ensemble deep learning architecture is considered that consists of Long Short-Term Memory and Gated Recurrent Unit models to capture dynamic temporal dependencies within emotional data sequences effectively. The publicly available wearable devices EMOGNITION database is utilized to facilitate result reproducibility and comparison. The database includes physiological signals recorded using the Samsung Galaxy Watch, Empatica E4 wristband, and MUSE 2 Electroencephalogram (EEG) headband devices for a comprehensive understanding of emotions. A detailed comparison of all three dedicated wearable devices has been carried out to identify nine discrete emotions, exploring three different bio-signal combinations. The Samsung Galaxy and MUSE 2 devices achieve an average classification accuracy of 99.14% and 99.41%, respectively. The performance of the Samsung Galaxy device is examined for the 2D Valence-Arousal effective dimensional model. Results reveal average classification accuracy of 97.81% and 72.94% for Valence and Arousal dimensions, respectively. The acquired results demonstrate promising outcomes in emotion recognition when compared with the state-of-the-art methods.
... The first part of the Individual leg test, named the Formula part, aimed to validate the correlation between grouser density and grouser height, as described by the equations in [5]. However, the results, shown in Table I along with a 95% confidence interval [7], revealed a significant disparity in friction among different angles. Consequently, the formula derived from this correlation cannot be applied in this scenario. ...
Experiment Findings
Full-text available
The Lunar Zebro project aims to develop an efficient robotic system capable of traversing the lunar surface and collecting valuable data. However, there exists a research gap regarding the optimal leg design for the Lunar Zebro and the process of selecting the most suitable legs. This paper aims to address a portion of this research gap by conducting individual tests on various grouser designs and investigating the impact of different grousers on the overall performance of the Lunar Zebro. Subsequently, the entire leg configuration with the optimal legs was tested against the previously used legs. Finally, the shapes of the legs will be varied systematically to discover the optimal configuration for the Lunar Zebro. This led to the conclusion that the combination of the different variables has significant implications for performance.
... The individual effects of six variables on the mentioned dependent variable were examined. A 99% confidence interval was used for evaluation [32]. In order to ensure the reliability of the results and due to the sample size, the partial influence of six variables on the dependent variable, the impact of branding activities on the company results, was tested using simple regression. ...
Article
Full-text available
In the face of a competitive and fast-evolving marketplace, companies are seeking new ways to gain a sustainable advantage and meet increasingly sophisticated consumer demands. This paper focuses on long-term brand development as a strategic investment for building consumer loyalty and maintaining market relevance. It provides a theoretical overview of branding by tracing its historical evolution and defining it as a modern strategic tool. Central to this analysis is the brand management process, where a structured system of performance indicators has been created to systematically monitor brand health and effectiveness. Drawing from both prior research and the authors' professional experiences, the paper proposes a model for evaluating brand impact. This model was tested through an empirical study of leading trading companies in Montenegro, allowing researchers to quantify the influence of specific variables and assess their interrelationships. Statistical techniques, including descriptive statistics, correlation, and regression analysis, were used to validate the model, revealing significant relationships that affirm its practical utility. The developed model can also be applied in other research areas, which may concern some other environments.
... The prevalence and 95% confidence intervals (95%CI) for self-reported total violence and subtypes were calculated per the explanatory variables. Differences between groups without overlapping 95%CI were assessed as significant 25 . ...
Article
Full-text available
The prevalence and factors associated with self-reported elder abuse were estimated. This was a cross-sectional study with data from people aged 60 or over interviewed in the 2019 PNS. Prevalence and 95% confidence intervals (95%CI) of total violence and subtypes were calculated according to selected variables, with analysis of associations between total violence and explanatory variables using Poisson Regression. Of the elderly, 10.05% reported some violence, 11.09% were women and 8.69% were men. In women, violence was more prevalent: 60 to 69 (PR: 2.18; 95%CI: 1.51;314) and 70 to 79 years old (PR: 1.91; 95%CI: 1.31;2.77), divorced/separated (PR: 1.74; 95%CI: 1.35;2.24), resident in urban areas (PR: 1.49; 95%CI: 1.20;1.61), regular self-rated health (PR: 1.34; 95%CI: 1.09;1.65) and poor/very poor (PR: 2.04; 95%CI: 1.64;2.55), depressive symptoms (PR: 2.13; 95%CI: 1.72;2.63) and multimorbidity (PR: 1.21; 95%CI: 1.01;1.44). In men, the prevalence was higher: 60 to 69 years (PR: 1.79; 95%CI: 1.19;2.69), non-white race/skin color (PR: 1.44; 95%CI: 1.18;1.75), resident in urban areas (PR: 1.28; 95%CI: 1.02;1.61), regular self-assessment of health (PR: 1.36; 95%CI: 1.09;1.70) and depressive symptoms (PR: 2.57; 95%CI: 1.84;3.56).
Article
Background Nurse managers are essential in mitigating burnout among staff nurses; however, they are also susceptible to burnout due to overwhelming workloads and emotional exhaustion. This study examines the effectiveness of interventions designed to increase burnout awareness among nurse managers and promote proactive strategies to combat it. Aims The study aimed to enhance nurse managers' awareness of burnout using a technology-based solution and assess the impact of targeted interventions. Methods Data were collected over 8 months utilizing self-report assessment, reflective journaling, a pre–post burnout survey (MBI-GS), individual coaching, and postintervention focus groups. Results Postintervention analysis showed improvement across all MBI-GS subscales and a notable improvement in burnout awareness. Twenty-eight percent of nurse managers remained unaware or blocked postintervention, compared to 43% preintervention. Overall, burnout awareness improved by 21%. Linking Evidence to Action This study highlights the importance of burnout awareness for nurse managers, demonstrating that a technology-based solution, combined with targeted human-centered interventions, supports technology adoption, burnout awareness, and the development of adaptive behaviors.
Article
Background Electronic gambling machines (EGMs) are among the most harmful forms of gambling. Their widespread availability has been linked to increased gambling activity and a rise in harmful gambling. However, it remains unclear whether restricting the availability and accessibility of EGMs is effective in reducing gambling harms at the population level. We investigate those issues by leveraging the natural experiment of the region of Piedmont in northwest Italy, which in 2018 set limits on the location and operating hours of EGMs. Methods Data from the Gambling Adult Population Survey conducted in 2018 and 2022 were combined with information regarding the implementation of the policy in Piedmont municipalities and data on EGMs from the national monopoly holder. Our analysis uses a difference-in-difference design, which is appropriate given that only some municipalities in the region implemented the policy, each setting different operating time limits for EGMs. Sensitivity analyses were conducted to explore possible substitution effects towards bordering regions and online gambling. Results Our findings reveal that policies aimed at reducing the operating hours of EGMs were effective in decreasing harmful gambling behaviour in the population, but only when a minimum daily operational suspension of 11 hours was implemented. No significant effect was observed from setting location limits, and no substitution effect was found. Conclusions While restricting EGM availability can help mitigate harmful gambling, it requires coordinated efforts across different levels of government to ensure uniform policy implementation, along with individual-level policies, such as banning advertising, and consideration of diverse gambling products and online gambling.
Article
Resumo Estimou-se a prevalência e os fatores associados ao autorrelato de VCPI. Estudo transversal com dados de pessoas com 60 anos ou mais entrevistados na PNS 2019. Calcularam-se prevalências e intervalos de confiança de 95% (IC95%) de violência total e subtipos, segundo variáveis selecionadas, com análise de associações entre violência total e variáveis explicativas pela Regressão de Poisson. Dos idosos, 10,05% relataram alguma violência, 11,09% mulheres e 8,69% homens. Nas mulheres, a violência foi mais prevalente: 60 a 69 (RP: 2,18; IC95%: 1,51;314) e 70 a 79 anos (RP: 1,91; IC95%: 1,31;2,77), divorciada/separada (RP: 1,74; IC95%: 1,35;2,24), residente em área urbana (RP: 1,49; IC95%: 1,20;1,61), autoavaliação de saúde regular (RP: 1,34; IC95%: 1,09;1,65) e ruim/muito ruim (RP: 2,04; IC95%: 1,64;2,55), sintomas depressivos (RP: 2,13; IC95%: 1,72;2,63) e multimorbidade (RP: 1,21; IC95: 1,01;1,44). Nos homens, a prevalência foi maior: 60 a 69 anos (RP: 1,79; IC95%: 1,19;2,69), raça/cor da pele não branca (RP: 1,44; IC95%: 1,18;1,75), residente em área urbana (RP: 1,28; IC95%: 1,02;1,61), autoavaliação regular de saúde (RP: 1,36; IC95%: 1,09;1,70) e sintomas depressivos (RP: 2,57; IC95%: 1,84;3,56).
Preprint
Full-text available
Multiple Sclerosis (MS) is a central nervous system (CNS) autoimmune inflammatory disease targeting the myelin sheath and affecting 2.8 million patients worldwide, mostly in economically advanced countries. The OFSEP-HD (French Multiple Sclerosis Registry - High Definition) multi-centric cohort comprises 2,667 genetic samples of patients with MS including 5 years of clinical, biological and imaging follow up. Here we described the genetic background of the cohort using data generated from the Affymetrix Precision Medicine Research Array (PMRA) genotyping chips to collect 888,799 genomic variants, and up to 8.5 million variants after imputation. Our analysis focused on genetic ancestry, admixture analysis and Human Leukocyte Antigen (HLA) including haplotypes inference. Principal Components Analysis (PCA) clustering identified seven ancestral clusters with 2177 patients (85.6 %) from clearly defined European ancestry. We observed 232 MS patients from North-African genetic ancestry while 120 of those patients (51.7%) did not self-report North-African origins, highlighting once again the limitations of self-assessed population descriptors. To promote data sharing we implemented the generation of a realistic and anonymous synthetic dataset using an adaptation of a known synthetic data generation methodology. This work unveils the genetic landscape and heterogeneous profiles of the OFSEP-HD cohort and proposes an open synthetic genetic dataset for further analyses.
Article
Full-text available
Application and interpretation of statistical evaluation of relationships is a necessary element in biomedical research. Statistical analyses rely on P value to demonstrate relationships. The traditional level of significance, P<0.05, can be negatively impacted by small sample size, bias, and random error, and has evolved to include interpretation of statistical trends, correction factors for multiple analyses, and acceptance of statistical significance for P>0.05 for complex relationships such as effect modification.
Article
Full-text available
Null hypothesis significance testing (NHST) is undoubtedly the most common inferential technique used to justify claims in the social sciences. However, even staunch defenders of NHST agree that its outcomes are often misinterpreted. Confidence intervals (CIs) have frequently been proposed as a more useful alternative to NHST, and their use is strongly encouraged in the APA Manual. Nevertheless, little is known about how researchers interpret CIs. In this study, 120 researchers and 442 students-all in the field of psychology-were asked to assess the truth value of six particular statements involving different interpretations of a CI. Although all six statements were false, both researchers and students endorsed, on average, more than three statements, indicating a gross misunderstanding of CIs. Self-declared experience with statistics was not related to researchers' performance, and, even more surprisingly, researchers hardly outperformed the students, even though the students had not received any education on statistical inference whatsoever. Our findings suggest that many researchers do not know the correct interpretation of a CI. The misunderstandings surrounding p-values and CIs are particularly unfortunate because they constitute the main tools by which psychologists draw conclusions from data.
Article
In radiology, appropriate diagnoses are often based on quantitative data. However, these data contain inherent variability. Radiologists often see P values in the literature but are less familiar with other ways of reporting statistics. Statistics such as the SD and standard error of the mean (SEM) are commonly used in radiology, whereas the CI is not often used. Because the SEM is smaller than the SD, it is often inappropriately used in order to make the variability of the data look tighter. However, unlike the SD, which quantifies the variability of the actual data for a single sample, the SEM represents the precision for an estimated mean of a general population taken from many sample means. Since readers are usually interested in knowing about the variability of the single sample, the SD often is the preferred statistic. Statistical calculations combine sample size and variability (ie, the SD) to generate a CI for a population proportion or population mean. CIs enable researchers to estimate population values without having data from all members of the population. In most cases, CIs are based on a 95% confidence level. The advantage of CIs over significance tests (P values) is that the CIs shift the interpretation from a qualitative judgment about the role of chance to a quantitative estimation of the biologic measure of effect. Proper understanding and use of these fundamental statistics and their calculations will allow more reliable analysis, interpretation, and communication of clinical information among health care providers and between these providers and their patients.
Article
The estimation approach to statistical analysis aims to quantify the effect of interest as an "estimate" of a clinically relevant quantity and to quantify the uncertainty in this estimate by means of a confidence interval (CI). As such, results expressed in this form are much more informative than results presented just as p values. This article focuses on the principles rather than the mathematics of CIs and discusses interpretation of CIs and some common misuses. CIs can be constructed for almost all analyses. They are especially useful for avoiding misinterpretation of nonsignificant results of small studies. CIs should be provided routinely for the main results of trials and observational studies.
Article
Clinical trials are usually performed on a sample of people drawn from the population of interest. The results of a trial are, therefore, estimates of what might happen if the treatment were to be given to the entire population of interest. Confidence intervals (CIs) provide a range of plausible values for a population parameter and give an idea about how precise the measured treatment effect is. CIs may also provide some useful information on the clinical importance of results and, like p-values, may also be used to assess ‘statistical significance’. Although other CIs can be calculated, the 95% CI is usually reported in the medical literature. In the long run, the 95% CI of an estimate is the range within which we are 95% certain that the true population parameter will lie. Despite the usefulness of the CI approach, hypothesis testing and the generation of p-values are common in the medical literature. The p-value is often used to express the probability that the observed differences between study groups are due to chance. p-values provide no information on the clinical importance of results. Conclusion: It is good practice for authors of research articles to report CIs with their estimates instead of just p-values as p-values are less informative and convey no information on clinical importance.
Article
When the aim of the randomized controlled trial (RCT) is to show that one treatment is superior to another, a statistical test is employed and the trial (test) is called a superiority trial (test). Often a nonsignificant superiority test is wrongly interpreted as proof of no difference between the two treatments. Proving that two treatments are equal in performance is impossible with statistical tools; at most, one can show that they are equivalent. In an equivalence trial, the statistical test aims at showing that two treatments are not too different in characteristics, where "not too different" is defined in a clinical manner. Finally, in a non-inferiority trial, the aim is to show that an experimental treatment is not (much) worse than a standard treatment. In this report, the three types of trials are compared, but the main focus is on the non-inferiority trial. Special attention is paid to the practical implications when setting up a non-inferiority trial. Illustrations are taken from a clinical trial in osteoarthritis and from thrombolytic research.
Confidence interval for a median and other quartiles
  • M Bland
Bland M. Confidence interval for a median and other quartiles [Monograph on the internet]. Available online: https://www-users.york.ac.uk/~mb55/intro/cicent.htm
What is a confidence interval and why would you want one?
  • Anonymous
Anonymous. What is a confidence interval and why would you want one? [Monograph on the internet]. Available online: http://www.uxmatters.com/mt/archives/2011/11/what-is-aconfidence-interval-and-why-would-you-want-one.php
Confidence intervals vs Bayesian intervals Foundations of probability theory, statistical inference, and statistical theories of science
  • Et Jaynes
Jaynes ET. Confidence intervals vs Bayesian intervals. In: Harper WL, Hooker CA, editors. Foundations of probability theory, statistical inference, and statistical theories of science. Vol II. Dordrecht: D. Reidel Publishing Company 1976:175-257.