<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Violation of Independence of Irrelevant Alternatives in Friedman's test</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jan Motl</string-name>
          <email>jan.motl@fit.cvut.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavel Kordík</string-name>
          <email>pavel.kordik@fit.cvut.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Czech Technical University in Prague</institution>
          ,
          <addr-line>Thákurova 9, 160 00 Praha 6</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>2203</volume>
      <fpage>59</fpage>
      <lpage>63</lpage>
      <abstract>
        <p>One of the most common methods for classifier comparison is Friedman's test. However, Friedman's test has a known flaw - ranking of classifiers A and B does not depend only on the properties of classifiersA and B, but also on the properties of all other evaluated classifiers. We illustrate the issue on a question: “What is better, bagging or boosting?”. With Friedman's test, the answer depends on the presence/absence of irrelevant classifiers in the experiment. Based on the application of Friedman's test on an experiment with 179 classifiers and 121 datasets we conclude that it is very easy to game the ranking of two insignificantly different classifiers. But once the difference becomes significant, it is unlikely that by removing irrelevant classifiers we obtain a significantly different classifiers but with reversed conclusion. Friedman's test Friedman's test, if applied on algorithm ranking, is defined as:</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Friedman’s test is the recommended way how to compare
algorithms in machine learning (ML) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. It is a
nonparametric test, which calculates scores not on the raw
performance measures (e.g. classification accuracy or correlation
coefficient) but on the ranks calculated from the raw
measures. Nonparametric tests are favored over parametric tests
in ML, because the standard formulation of the central limit
theorem (CLT) does not apply on bounded measures [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ],
many measures used in ML are bounded, and commonly
used parametric tests rely on CLT. Nevertheless, even
nonparametric tests, which are based on ranking, have flaws
as demonstrated by Arrow’s impossibility theorem [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In
this article, we discuss one such flaw of Friedman’s test:
violation of Independence of Irrelevant Alternatives (IIA).
      </p>
      <p>Given data {xi j}n×k, that is, a matrix with n rows (the
datasets), k columns (the algorithms) and a single
performance observation at the intersection of each
dataset and algorithm, calculate the ranks within each
dataset. If there are tied values, assign to each tied
value the average of the ranks that would have been
assigned without ties. Replace the data with a new
matrix {ri j}n×k where the entry ri j is the rank of xi j
within dataset i. Calculate then rank sums of
algon
rithms as: r j = ∑i=1 ri j.</p>
      <p>
        We can rank the algorithms based on rank sums r j [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
Friedman’s test then continues with the evaluation of the
null hypothesis that there are no differences between the
classifiers. Since this article is not about hypothesis testing
but rather about algorithm ranking based on r j, we
reference a keen reader to read [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for a detailed description of
Friedman’s test hypothesis testing.
      </p>
      <p>
        IIA Independence of Irrelevant Alternatives [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] condition
is defined as:
      </p>
      <p>If algorithm A is preferred to algorithm B out of the
choice set {A,B}, introducing a third option X ,
expanding the choice set to {A,B,X }, must not make B
preferable to A.</p>
      <p>In other words, preferences for algorithm A or algorithm
B, as determined by rank sums r j, should not be changed
by the inclusion of algorithm X , i.e., X is irrelevant to the
choice between A and B.</p>
      <p>Illustration Boosting tends to outperform base algorithms
(e.g. decision trees) by a large margin. But sometimes,
boosting fails [2, Chapter 8.2] while bagging reliably
outperforms base algorithms on all datasets, even if only by a
small margin [2, Chapter 7.1]. This is illustrated in the left
part of Figure 1. If we compare boosting only to bagging,
then by the rank sums r j, boosting wins, because boosting
is better than bagging on the majority of datasets. However,
if we add irrelevant algorithms that are always worse than
bagging, the conclusion may change. Bagging will be
always the first or the second. But boosting will be either the
first or (in the provided illustration) the last. And a few
extreme values in the rank sum r j can result into the change
of the conclusion.
t
e
s
a
t
a
D
1-p {</p>
    </sec>
    <sec id="sec-2">
      <title>Measurements (bigger is better)</title>
      <p>m Irrelevant Bagging Boosting
...</p>
      <p>...</p>
      <p>Measure (e.g. AUC)</p>
      <p>Ranks (bigger is better)
m Irrelevant Bagging Boosting
. . . m+1 m+2
. . . m+1 m+2
... .. . m+...1 m+... 2
. . . m+1 m+2
... .. . m+...2 1...</p>
      <p>. . . m+2 1
Avg rank rj (1+-pp()m(m++21)) (1-p)(m+2)
+ p
Equilibrium</p>
      <p>rBagging = rBoosting
(1-p)(m+1)+p(m+2) = (1-p)(m+2)+p
p(m+2) = 1</p>
    </sec>
    <sec id="sec-3">
      <title>When p and/or m increase, bagging wins over boosting</title>
      <p>Hypothesis Authors of highly cited papers, knowingly or
not, frame their algorithms in the best possible light. Based
on the equilibrium equation in Figure 1, we would
expect that proponents of boosting algorithms use fewer
algorithms than proponents of bagging in the design of
experiments (DOE).</p>
      <p>
        Evaluation Breiman, the author of bagging [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], compared
bagging against 22 other algorithms while Freund and
Schapire, authors of AdaBoost [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], compared their
boosting algorithm against only 2 other algorithms. Our
expectations were fulfilled.
      </p>
      <p>Threats to validity due to omitted variables Since both
articles were published in the same year, the availability
of algorithms for comparison should be comparable and
cannot be used to explain the observed differences in the
DOE.</p>
      <p>Conclusion Since this is just a single observation, which
can be just by chance, following section analyses the
impact of DOE numerically.
3.2</p>
      <sec id="sec-3-1">
        <title>Effect on Large Studies</title>
        <p>
          A nice comparison of 179 classifiers on 121 datasets is
provided by Fernández-Delgado et al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. The authors
follow Demšar’s [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] recommendation to use Friedman’s
test to rank the algorithms.
        </p>
        <p>If we directly compare just two classifiers,
AdaBoostM1_J48_weka and Bagging_J48_weka, and
calculate rank sums r j, then we get that boosting beats bagging
64 to 47. But if we calculate rank sums over all algorithms,
then we get that bagging beats boosting 13136 to 12898.
A completely reversed conclusion!
How frequently does the ordering flip? If we
perform above analysis over all unique pairs of classifiers
( 12 179(179 − 1) = 15753), then we get that the ordering
flips in 5% of cases (831/15753).</p>
        <p>Are the changes in the ranks significant? We repeated the
experiment once again, but considered only classifier pairs
that are based on Friedman’s test:</p>
        <sec id="sec-3-1-1">
          <title>1. pairwise significantly different</title>
          <p>2. and significantly different in the presence of all the
remaining classifiers .</p>
          <p>
            If we do not apply multiple testing correction, the
classifiers flip the order in mere 0.05% cases (8/15753) at a
shared significance level α = 0.05. Once we add
multiple testing correction, the count of significant flips drops
to zero. In our case, the exact type of multiple testing
correction does not make a difference. Bonferroni, Nemeneyi,
Finner &amp; Li and Bergmann &amp; Hommel [
            <xref ref-type="bibr" rid="ref17">17</xref>
            ] corrections all
give the same result as the lowest p-value before the
correction in that 8 cases is 0.023, which is already extremely
close to the significance level of 0.05.
          </p>
          <p>
            Related Work and Discussion
We are not the first one to notice that Friedman’s test does
not fulfill the IIA property. The oldest mention of the issue
in the literature, that we were able to track down, dates
back to 1966 [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]. Following paragraphs discuss possible
solutions to the issue.
4.1
          </p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Pairwise Ranking</title>
        <p>
          Benevoli et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] recommend replacing Friedman’s test
with pairwise Wilcoxon signed-rank test followed with
multiple testing correction. The application of a pairwise-test
solves the issue with IIA if we want to show how well “a
new algorithm A fares in comparison to a set of old
algorithms B” because each pairwise comparison between
A and some other algorithm B ∈ B is by definition
independent on C ∈ B,C 6= B. However, if we aim to “rank the
current algorithms together”, pairwise tests may not deliver
a total ordering while a method comparing all algorithms
at once can, as illustrated in Figure 2.
        </p>
        <p>α
β
γ
δ
ε</p>
        <sec id="sec-3-2-1">
          <title>Rank sum 10</title>
          <p>
            A
to come with a fair vote method arose. Two competitors,
Condorcet and Borda1, came with two different methods.
And they did not manage to agree, which of the methods
is better. This disagreement spurred an interest into vote
theory. One of the possible alternatives to Friedman test
that vote theory offers is Ranked pairs method [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ]:
1. Get the count of wins for each pair of
algo
          </p>
          <p>rithms.
2. Sort the pairs by the difference of the win counts</p>
          <p>in the pair.
3. “Lock in” the pairs beginning with the strongest</p>
          <p>difference of the win counts.</p>
          <p>While Ranked pairs method also fails IIA criterium, it at
least fulfills a weaker criterium calledLocal Independence
from Irrelevant Alternatives (LIIA). LIIA requires that the
following conditions hold:</p>
          <p>If the best algorithm is removed, the order of the
remaining algorithms must not change. If the worst
algorithm is removed, the order of the remaining
algorithms must not change.</p>
          <p>
            An example where Friedman’s test fails LIIA criterium
is given in Figure 3 [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ].
          </p>
          <p>α
β
γ
δ
ε
Rank sum</p>
          <p>A
Ak. When A0 is better, the rank is A0 A1 . . .</p>
          <p>Ak B. Therefore, the presence of A1, A2, . . . , Ak
artificially increases the difference betweenA0 and B.</p>
          <p>But Ranked pairs method fulfills this criterion.</p>
          <p>
            Finally, Friedman’s test violates Majority criterion [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ]:
If one algorithm is the best on more than 50% of
datasets, then that algorithm must win.
          </p>
          <p>
            We have already observed a violation of this criterion in
the though example with bagging versus boosting — even
thought boosting was the best algorithm on the majority
of the datasets, this fact alone did not guarantee boosting’s
victory. The violation of Majority criterion also implies a
violation of Condorcet criterion [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ]:
          </p>
          <p>If there is an algorithm which wins pairwise to each
other algorithm, then that algorithm must win.</p>
          <p>
            Which is, nevertheless, once again fulfilled with Ranked
pairs method. However, just like in machine learning we
have no free lunch theorem, Arrow’s impossibility
theorem [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] states that there is not a ranked vote method
without a flaw. The flaw of all ranked vote methods, but
dictatorship2, that fulfill Condorcet criterium is that they fail
Consistency criterium [18, Theorem 2]:
          </p>
          <p>If based on a set of datasets A an algorithm A is the
best. And based on another set of datasets B an
algorithm A is, again, the best. Then based on A ∪ B the
algorithm A must be the best.</p>
          <p>Notably, Friedman’s test fulfills this criterium while
Ranked pairs method fails this criterium. For convenience,
a summary table with the list of discussed criteria is given
in Table 1.
Contrary to our expectations, based on analysis of 179
classifiers on 121 datasets, Friedman’s test appears to be fairly
resistant to manipulation, where we add or remove
irrelevant classifiers from the analysis. Therefore, wecannot
recommend avoiding Friedman’s test only because it violates
Independence of Irrelevant Alternatives (IIA) criterium.
6</p>
          <p>Acknowledgment
I would like to thank Adéla Chodounská for her help. We
furthermore thank the anonymous reviewers, their
comments helped to improve this paper. The reported research
has been supported by the Grant Agency of the Czech
Technical University in Prague (SGS17/210/OHK3/3T/18) and
the Czech Science Foundation (GACˇ R 18-18080S).</p>
          <p>Non-dictatorship The outcome should not depend only
upon a single dataset.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Independence of Irrelevant Alternatives (IIA) If algo</title>
        <p>rithm A is preferred to algorithm B out of the choice
set {A,B}, introducing a third option X , expanding the
choice set to {A,B,X }, must not make B preferable to
A.</p>
        <p>Appendix: Criteria</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Kenneth</surname>
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Arrow</surname>
          </string-name>
          .
          <article-title>A Difficulty in the Concept of Social Welfare</article-title>
          . J.
          <string-name>
            <surname>Polit</surname>
          </string-name>
          . Econ.,
          <volume>58</volume>
          (
          <issue>4</issue>
          ):
          <fpage>328</fpage>
          -
          <lpage>346</lpage>
          ,
          <year>1950</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Eric</given-names>
            <surname>Bauer</surname>
          </string-name>
          and
          <string-name>
            <given-names>Ron</given-names>
            <surname>Kohavi</surname>
          </string-name>
          .
          <article-title>An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and</article-title>
          <string-name>
            <surname>Variants. Mach. Learn.</surname>
          </string-name>
          ,
          <volume>36</volume>
          (
          <issue>1-2</issue>
          ):
          <fpage>105</fpage>
          -
          <lpage>139</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Harry</given-names>
            <surname>Beatty</surname>
          </string-name>
          .
          <article-title>Voting Rules and Coordination Problems</article-title>
          . In Methodol. Unity Sci., pages
          <fpage>155</fpage>
          -
          <lpage>189</lpage>
          . Springer Netherlands, Dordrecht,
          <year>1973</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Alessio</given-names>
            <surname>Benavoli</surname>
          </string-name>
          , Giorgio Corani, and
          <string-name>
            <given-names>Francesca</given-names>
            <surname>Mangili</surname>
          </string-name>
          .
          <article-title>Should we really use post-hoc tests based on mean-ranks?</article-title>
          <source>J. Mach. Learn. Res.</source>
          ,
          <volume>17</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Duncan</given-names>
            <surname>Black</surname>
          </string-name>
          .
          <article-title>On the Rationale of Group Decision-making</article-title>
          .
          <source>J. Polit. Econ.</source>
          ,
          <volume>56</volume>
          (
          <issue>1</issue>
          ):
          <fpage>23</fpage>
          -
          <lpage>34</lpage>
          , feb
          <year>1948</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Leo</given-names>
            <surname>Breiman</surname>
          </string-name>
          .
          <article-title>Bagging predictors</article-title>
          . Mach. Learn.,
          <volume>24</volume>
          (
          <issue>2</issue>
          ):
          <fpage>123</fpage>
          -
          <lpage>140</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Giorgio</given-names>
            <surname>Corani</surname>
          </string-name>
          , Alessio Benavoli, Janez Demšar, Francesca Mangili, and
          <string-name>
            <given-names>Marco</given-names>
            <surname>Zaffalon</surname>
          </string-name>
          .
          <article-title>Statistical comparison of classifiers through Bayesian hierarchical modelling</article-title>
          .
          <source>Mach</source>
          . Learn.,
          <volume>106</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1817</fpage>
          -
          <lpage>1837</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Nicholas</surname>
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Cox</surname>
          </string-name>
          .
          <article-title>Correlation with confidence, or Fisher's z revisited</article-title>
          .
          <source>Stata J.</source>
          ,
          <volume>8</volume>
          (
          <issue>3</issue>
          ):
          <fpage>413</fpage>
          -
          <lpage>439</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Janez</given-names>
            <surname>Demšar</surname>
          </string-name>
          .
          <article-title>Statistical Comparisons of Classifiers over Multiple Data Sets</article-title>
          .
          <source>J. Mach. Learn. Res.</source>
          ,
          <volume>7</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>30</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Manuel</given-names>
            <surname>Fernández-Delgado</surname>
          </string-name>
          , Eva Cernadas, Senén Barro, and
          <string-name>
            <given-names>Dinani</given-names>
            <surname>Amorim</surname>
          </string-name>
          .
          <article-title>Do we Need Hundreds of Classifiers to Solve Real World Classification Problems?</article-title>
          J.
          <string-name>
            <surname>Mach</surname>
          </string-name>
          .
          <source>Learn. Res.</source>
          ,
          <volume>15</volume>
          :
          <fpage>3133</fpage>
          -
          <lpage>3181</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Yoav</given-names>
            <surname>Freund</surname>
          </string-name>
          and
          <string-name>
            <given-names>Robert E.</given-names>
            <surname>Schapire</surname>
          </string-name>
          .
          <article-title>Experiments with a New Boosting Algorithm</article-title>
          .
          <source>Int. Conf. Mach. Learn.</source>
          , pages
          <fpage>148</fpage>
          -
          <lpage>156</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>John</surname>
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschke</surname>
          </string-name>
          .
          <article-title>Bayesian data analysis</article-title>
          .
          <source>Wiley Interdiscip. Rev. Cogn. Sci.</source>
          ,
          <volume>1</volume>
          (
          <issue>5</issue>
          ):
          <fpage>658</fpage>
          -
          <lpage>676</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>John</surname>
            <given-names>W. Pratt.</given-names>
          </string-name>
          <article-title>Remarks on Zeros and Ties in the Wilcoxon Signed Rank Procedures</article-title>
          .
          <source>Am. Stat. Assoc.</source>
          ,
          <volume>54</volume>
          (
          <issue>287</issue>
          ):
          <fpage>655</fpage>
          -
          <lpage>667</lpage>
          ,
          <year>1959</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Paramesh</given-names>
            <surname>Ray</surname>
          </string-name>
          .
          <source>Independence of Irrelevant Alternatives. Econometrica</source>
          ,
          <volume>41</volume>
          (
          <issue>5</issue>
          ):987, sep
          <year>1973</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Markus</given-names>
            <surname>Schulze</surname>
          </string-name>
          .
          <article-title>A new monotonic and clone-independent single-winner election method</article-title>
          .
          <source>Voting Matters</source>
          , (
          <volume>17</volume>
          ):
          <fpage>9</fpage>
          -
          <lpage>19</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Thorwald</surname>
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tideman</surname>
          </string-name>
          .
          <article-title>Independence of clones as a criterion for voting rules</article-title>
          .
          <source>Soc. Choice Welfare</source>
          ,
          <volume>4</volume>
          (
          <issue>3</issue>
          ):
          <fpage>185</fpage>
          -
          <lpage>206</lpage>
          ,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17] Bogdan Trawin´ski, Magdalena Smetek, Zbigniew Telec, and
          <string-name>
            <given-names>Tadeusz</given-names>
            <surname>Lasota</surname>
          </string-name>
          .
          <article-title>Nonparametric statistical analysis for multiple comparison of machine learning regression algorithms</article-title>
          .
          <source>Int. J. Appl. Math. Comput. Sci.</source>
          ,
          <volume>22</volume>
          (
          <issue>4</issue>
          ):
          <fpage>867</fpage>
          -
          <lpage>881</lpage>
          , jan
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Young</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Levenglick</surname>
          </string-name>
          .
          <article-title>A Consistent Extension of Condorcet's Election Principle</article-title>
          .
          <source>SIAM J. Appl</source>
          . Math.,
          <volume>35</volume>
          (
          <issue>2</issue>
          ):
          <fpage>285</fpage>
          -
          <lpage>300</lpage>
          , sep
          <year>1978</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>