ArticlePDF Available

Response to van Pul, van Zanten and Wichink Kruit

Authors:
  • University College Roosevelt (UU)
  • Clintel Foundation
LETTER TO THE EDITOR
Response to van Pul, van Zanten and Wichink Kruit
Pul et al. offer a critique of our recent contribution in Soil
Use and Management. Regrettably, their critique contains a
number of confusing statements and unstated assumptions,
some of which we identify here.
Van Pul et al. state that the ‘statistical correlation between
hourly ammonia concentrations between regional
measurement stations is weak’, which, as we pointed out, is
true. They ascribe this weakness to ‘large variability in local
agricultural practice and in weather conditions’, which is
also true. They then say: ‘If data are aggregated to longer
timescales, correlations between stations clearly increase’,
which happens, but they claim this increased correlation is
‘due to the removal of noise at the hourly timescale’.
First, the ‘removal of noise’ logically assumes that their
statistical model is more real than the measured values. By
implication, the measured values seem to be viewed by Van Pul
et al. as a corrupted form of reality in the sense there is a ‘true’
value to which ‘noise’ is added, and that their statistical model
can discover this ‘true’ value. This clearly is false. Instead, the
sum total of all causes results in certain ammonia concentration
at time x and location y and recorded as nlg/m
3
. These
measured values are the experienced values; that is, what is
experienced is not some mysterious truth after which noise has
somehow been subtracted. The subsequent measured data are
the actual values. As we made clear in our study, no
assumptions are made by us concerning the data from the
Dutch National Air Quality Monitoring Network (LML).
This brings us to our second point: that data aggregation
increases correlation. What Van Pul et al. ignore, and we
pointed out (Hanekamp et al., 2017, p. 284), is that ‘smoothing
the data’ does not entail increased causation. As an experiment,
take any two sets of numbers, wholly unrelated to each other,
sampled at some finite timescale. Clearly, the correlation
between the two series will be quite small. Then, aggregate the
series by averaging (at least two) time points. The correlation
will, on average, increase (i.e. the absolute value). Aggregate
again at even coarser levels and the correlation increases once
more and may reach ‘significance’. Aggregation is a form of
statistical smoothing, and smoothing any two (or more) series
increases their correlation, on average, as Briggs (2016, p. 242
243) formally proved. Thus, how much of the correlation in the
ammonia series Van Pul et al. report is due to the artefacts of
smoothing, and how much is capturing larger-scale causes is
unknown. Yet, Van Pul et al. simply assume, without proof,
the latter to be the dominant, or indeed the only source of the
increased correlation.
Van Pul et al. next state that ‘annual concentrations at the
various stations over the period 19932014 tend to be
remarkably similar when normalized for the station’s
concentration level.’ By implication, they were not similar
until after statistical manipulation. There are, of course,
models that help characterize uncertainty in heterogeneous
time series like this, such as hierarchical modelling, which
might suggest larger-scale correlations, but, of course, never
causes. But even without these, it is obvious from the data
themselves that local causes dominate larger, more regional
causes. Otherwise, there would no need for statistical
manipulation.
We reiterate that the mean is an inadequate value to
represent average exposure. Indeed, ecosystems do not
experience average exposures at all, but a continuum of
differing concentrations in time, of which the high values are
short-lived, as is clear from the LML-data. Indubitably, it
would be best not to use point measures at all for complex
phenomena and use distributions. If the interest were in total
deposition, then it would be better to form a real measure of
total deposition and not use a proxy that gives undue weight
to temporarily high values, which does not represent average
system behaviour (Galton, 1907).
Van Pul et al. claim ‘annual concentrations of ammonia ...
appear to be only piecewise linear’ over various times.
They then suggest that ‘most stations have the largest
positive trends somewhere after the year 2000.’ and indicate
ways of statistically handling these pieces. This we consider
to be their greatest mistake. First, the periods Van Pul et al.
chose were biased and not with respect to known causes.
Anyone can pick arbitrary periods. Second, statistical models
are not needed to say whether a trend was present or absent
in any set of data. All one needs is a definition of ‘trend’
and then just observe the data. Does ‘trend’ mean more ‘ups’
than ‘downs’, or greater averages at the end than the
beginning, or so many per cent more ‘ups’, or higher or
lower at the end? Many more definitions of trends can be
proffered. Even if the definition is the trend coefficient of
(say) a linear regression model, then whether that ‘trend’ is
‘significant’ carries no meaning. It is certainly not an
indication that some linear cause was in effect, as the data
themselves by definition never harbour causal relations
(Briggs, 2014, 2016). No: something causes each data point,
and if the causes are unknown, which clearly is the case
here, then, statistical models are only worthwhile for making
predictions of what has not happened yet (Briggs, 2016).
Put differently: What makes the time periods noted by
Van Pul et al. the ‘correct’ periods? They simply observed
these periods after all. Why not, say, 2001 to 2002, or 2002
to 2003? Was there a trend in any of these series?
©2017 British Society of Soil Science 609
Soil Use and Management, December 2017, 33, 609–610 doi: 10.1111/sum.12385
SoilUse
and Management
Absolutely, and everywhere: no data series with real-world
causes remain constant. If the period of one year is
considered too short: why? After all, something (actually
many things) causes the series to change values. By seemly
and biased choice of time periods, trends have been
ostensibly demonstrated by Van Pul et al. and, in so doing,
beg the question about the cause.
If some large-scale, multi-year and linear cause is suspected,
then we first have to guess or deduce how that cause will affect
the series at hand, then verify if the data are indicative of that
cause. That is not the case here. If some cause is suspected but
its nature cannot be deduced, then various statistical models
can be posited to make predictions. If a strictly linear forcing,
for example, is suspected, then probabilistic predictions can be
made. This van Pul et al. have never done; they simply assume
the large-scale, multi-year and linear cause, namely
agricultural activities, and once again beg the question.
Considering the above, Van Pul et al. did not produce a case
against our critique. We therefore invite Van Pul et al. to do
the concrete predictive work.
W. M. BRIGGS
1
,J.C.HANEKAMP
2,3
&M.CROK
4
1
Independent researcher,
2
University College Roosevelt, Middelburg, The Netherlands
3
Environmental Health Sciences, University of
Massachusetts, Amherst, MA, USA
4
Independent researcher
E-mails: matt@wmbriggs.com; j.hanekamp@ucr.nl;
hjaap@xs4all.nl; marcel.crok@gmail.com
References
Briggs, W.M. 2014. Common statistical fallacies. Journal of
American Physicians and Surgeons,19,5860.
Briggs, W.M. 2016. Uncertainty The soul of modeling. Probability
& Statistics, Springer, Switzerland.
Galton, F. 1907. One vote. One value. Nature,75, 414.
Hanekamp, J.C., Briggs, W.M. & Crok, M. 2017. A volatile
discourse reviewing aspects of ammonia emissions, models and
atmospheric concentrations in The Netherlands. Soil Use and
Management,33, 276287.
©2017 British Society of Soil Science, Soil Use and Management,33, 609–610
610 Letter to the Editor
Technical Report
Full-text available
De metingen van het belangrijkste agrarische ammoniak meetstation (RIVM-meetstation Vredepeel op het terrein van Wageningen UR Plant Research) in het Landelijk Meetnet Luchtkwaliteit (LML) worden vervuild door een directe lokale bron, een pluimveehouderij, die te dicht op het meetstation staat.
Article
Full-text available
In the Netherlands, there is a vigorous debate on ammonia emissions, atmospheric concentrations and deposition between stakeholders and research institutions. In this article, we scrutinise some aspects of the ammonia discourse. In particular, we want to improve the understanding of the methodology for handling experimentally determined ammonia emissions. We show that uncertainty in published results is substantial. This uncertainty is under- or even unreported, and as a result, data in national emission inventories are overconfident by a wide margin. Next, we demonstrate that the statistical handling of data on atmospheric ammonia concentrations to produce national yearly atmospheric averages is oversimplified and consequently atmospheric concentrations are substantially overestimated. Finally, we show that the much-discussed ‘ammonia gap’ – either the discrepancy between calculated and measured atmospheric ammonia concentrations or the difference observed between estimated NH3 emission levels and those indicated by atmospheric measurements – is an expression of the widespread overconfidence placed in atmospheric modelling.
Article
A CERTAIN class of problems do not as yet appear to be solved according to scientific rules, though they are of much importance and of frequent recurrence. Two examples will suffice. (1) A jury has to assess damages. 2) The council of a society has to fix on a sum of money, suitable for some particular purpose. Each voter, whether of the jury or of the council, has equal authority with each of his colleagues. How can the right conclusion be reached, considering that there may be as many different estimates as there are members? That conclusion is clearly not the average of all the estimates, which would give a voting power to ``cranks'' in proportion to their crankiness. One absurdly large or small estimate would leave a greater impress on the result than one of reasonable amount, and the more an estimate diverges from the bulk of the rest, the more influence would it exert. I wish to point out that the estimate to which least objection can be raised is the middlemost estimate, the number of votes that it is too high being exactly balanced by the number of votes that it is too low. Every other estimate is condemned by a majority of voters as being either too high or too low, the middlemost alone escaping this condemnation. The number of voters may be odd or even. If odd, there is one middlemost value; thus in 11 votes the middlemost is the 6th; in 99 votes the middlemost is the 50th. If the number of voters be even, there are two middlemost values, the mean of which must be taken; thus in 12 votes the middlemost lies between the 6th and the 7th; in 100 votes between the 50th and the 51st. Generally, in 2n-1 votes the middlemost is the nth; in 2n votes it lies between the nth and the (n + 1)th.
Common statistical fallacies
  • Briggs W.M.
Briggs, W.M. 2014. Common statistical fallacies. Journal of American Physicians and Surgeons, 19, 58-60.
Uncertainty -The soul of modeling
  • W M Briggs
Briggs, W.M. 2016. Uncertainty -The soul of modeling. Probability & Statistics, Springer, Switzerland.