<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Predicting responses of individual reasoners in syllogistic reasoning by using collaborative filtering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ilir Kola</string-name>
          <email>kola@informatik.uni-freiburg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Ragni</string-name>
          <email>ragni@informatik.uni-freiburg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cognitive Computation Lab, University of Freiburg</institution>
          ,
          <addr-line>79110 Freiburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A syllogism consists of two premises each containing one of four quantifiers (All, Some, Some not, None) and two out of three objects totaling in 64 reasoning problems. The task of the participants is to draw or evaluate a conclusion, given the premise information. Most, if not all cognitive theories for syllogistic reasoning, focus on explaining and sometimes predicting the aggregated response pattern for participants of a whole psychological experiment. While only few theories focus on the level of an individual reasoner that might have a specific mental representation that explains her response pattern. If different reasoners can be grouped into similar answer patterns then it is possible to identify even cognitive styles that depend on the underlying representation. To test the idea of individual predictions, we start by developing a pair-wise similarity function based on the subjects' answers to the task. For 10% of the subjects, we randomly delete 15% of their answers. By using collaborative filtering techniques, we check whether it is possible to predict the deleted answers of a specific individual solely by using the answers given by similar subjects to those specific questions. Results show that not only the correct answer is predicted in around 70% of the cases, and the answer is in the top two predictions in 89% of the cases, which outperforms other theoretical approaches, but the predictions are as well accurate for cases where participants deviate from the correct answer. This implies that there are cognitive principles responsible for the patterns. If these principles are identified, then there is no need for complex models, because even simple ones can achieve high accuracy. This supports that individual performance in reasoning tasks can be predicted leading to a new level of cognitive modeling.</p>
      </abstract>
      <kwd-group>
        <kwd>computational reasoning</kwd>
        <kwd>individual differences</kwd>
        <kwd>syllogisms</kwd>
        <kwd>collaborative filtering</kwd>
        <kwd>machine learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        tions of the individual participants from the norms of classical logic have led to a
qualification of such reasoners as rather irrational (e.g., [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]). Another possibility is
that there is a so-called bounded rationality [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. An indicator could be that these
“deviators” are inherently consistent in their answers and even more that their answers
can be predicted. Most previous work has focused on overall distribution of answers,
trying to predict the most chosen answer by subjects. However, as noted by Pachur
and colleagues [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], in presence of individual differences, tests of group mean
differences can be highly misleading. For this reason, we focus on individual subjects and
try to predict the exact answer they would give.
      </p>
      <p>
        Collaborative filtering, a method employed in recommender systems [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], can
show that a single reasoner does not deviate from similar reasoners, and that
consequently her answers can be predicted based on answers of the similar reasoners.
      </p>
      <p>The rest of this paper is structured as follows: first, we give an introduction to
theories on reasoning and individual differences, syllogistic reasoning, and recommender
systems. Then, we present a model which uses collaborative filtering to predict
answers in the syllogistic reasoning task and compare it to other models or theoretical
predictions. Lastly, we draw conclusions and suggest further steps for research.</p>
    </sec>
    <sec id="sec-2">
      <title>Background</title>
      <sec id="sec-2-1">
        <title>Theories on reasoning and individual differences in reasoning</title>
        <p>
          Scientist have tried to understand human reasoning for a long time. Up to date, there
are at least five more prominent theories on how people reason. These theories are
based on heuristics [
          <xref ref-type="bibr" rid="ref14 ref3 ref4 ref6">3,4,6,14</xref>
          ] mental logic [
          <xref ref-type="bibr" rid="ref24 ref25">24,25</xref>
          ], pragmatic reasoning schemas [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ],
mental models [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], and probability theory [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Oaksford and Chater [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] offer a
general review of these theories.
        </p>
        <p>
          The need for all these theories is caused by the fact that people differ in how they
answer to reasoning tasks. Theories usually aim at explaining general answering
patterns, but if we focus on individual answers then these differences are even more vast.
These differences can be caused by intellectual abilities, memory capacity, strategies
being used, among others [
          <xref ref-type="bibr" rid="ref17 ref26">17, 26</xref>
          ].
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Syllogistic reasoning</title>
        <p>
          In a syllogistic task, subjects are presented with two premises, and they have to
evaluate what follows or whether a third given conclusion necessarily follows. Consider the
following example [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]:
Some Artists are Bakers,
All Bakers are Chemists.
        </p>
        <p>Therefore, some Artists are Chemists.</p>
        <p>Each premise can have four possible moods, two of which are affirmative (Some,
All), and two are negative ones (Some not, No). The premises have two terms each,
but overall only three terms are used. This is because the first two premises always
share a common term (in this case bakers), and the third premise asks about the
remaining two terms (artists and chemists). Terms can have four figures, based on their
configuration:</p>
        <p>Since each premise can have four moods, and there are four possible figures, there
can be 64 distinct pairs of premises. 27 of them have a conclusion which is valid in
classical logic, whereas for the remaining 37 there is no valid conclusion. The
conclusion (a third statement) allows again four possible moods, and two figures (A-C or
CA), so overall there are 512 syllogisms that can be evaluated.</p>
        <p>
          Studies using syllogisms with different forms of content from abstract to realistic
one have shown that errors are not random, but are systematically according to two
main factors: figure and mood (see [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]). Syllogistic reasoning has caught the attention
of many researchers. Khemlani and Johnson-Laird [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] provide a review of seven
theories of syllogistic reasoning. We will describe the ones which perform better in
the meta-analysis, and they will be later used as a baseline for the performance of our
model.
        </p>
        <p>
          The first theory, illicit conversions [
          <xref ref-type="bibr" rid="ref1 ref22">1,22</xref>
          ] is based on a misinterpretation of the
quantifiers interpreting All B are A when given All A are B and Some B are not A
when told Some A are not B. Both these conversions are logically invalid, and lead to
errors such as inferring All c are a given the premises All A are B and All C are B. In
order to predict the answers of syllogisms, this theory uses classical logic conversions
and operators, as well as the two aforementioned invalid conversions.
        </p>
        <p>
          The verbal models theory [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] claim that reasoners built verbal mental models
from syllogistic premises and either formulates a conclusion or declares that nothing
follows. The model then performs a reencoding of the information based on the
information that the converse of the quantifiers Some and No are valid. In another
version, the model also reencodes invalid conversions. The authors argue that a crucial
part of deduction is the linguistic process of encoding and reencoding the information,
rather than looking for counterexamples.
        </p>
        <p>
          Unlike the previous example, mental models (formulated for syllogisms first in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ])
are inspired by the use of counterexamples. The core idea is that individuals
understand that a putative conclusion is false if there is a counterexample to it. The theory
states that when being faced with a premise, individuals build a mental model of it
based on meaning and knowledge. E.g. when given the premise All Artists are
Beekeepers the following model is built:
        </p>
        <p>Each row represents the properties of an individual, and the ellipsis denotes
individuals which are not artists. This model can be fleshed out to an explicit model
which contains information on all potential individuals:</p>
        <sec id="sec-2-2-1">
          <title>Artist Beekeeper</title>
          <p>Artist Beekeeper</p>
          <p>Beekeeper</p>
          <p>In a nutshell, the theory states that many individuals simply reach a conclusion
based on the first implicit model, which can be wrong (in this case it would give the
impression that All Beekeepers are Artists). However, there are individuals who built
other alternative models in order to find counterexamples, which usually leads to a
logically correct answer.
2.3</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Collaborative filtering and recommender systems</title>
        <p>
          Recommender systems are software tools used to provide suggestions for items which
can be useful to users [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. One way to implement a recommender system is through
collaborative filtering. In a nutshell, collaborative filtering suggests that if Alice likes
items 1 and 2, and Bob likes items 1, 2 and 3, then Alice also probably likes item 3.
More formally, in collaborative filtering we look for patterns in observed preference
behavior, and try to predict new preferences based on those patterns. Users’
preferences for the items are stored as a matrix, in which each row represents a user and
each column represents an item. Then, for each user we build a similarity function to
see who are the users which have similar preferences. This means, for each user we
have a neighborhood of other users similar to them. Then, when a certain item has not
been rated by our user, we rely on this neighborhood to see how would our user rate
that item. If the rate would be high enough, we can recommend that item to the user.
        </p>
        <p>The main challenge in this case would be to select the appropriate similarity
function, and to determine the adequate size of the neighborhood.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Predicting performance in syllogistic reasoning by using collaborative filtering</title>
      <sec id="sec-3-1">
        <title>Motivation</title>
        <p>As aforementioned, people make mistakes when solving reasoning tasks such as
syllogisms. When it comes to preference behavior, we have seen that collaborative
filtering can achieve very good results in predicting which items to recommend to users.
This shows that people are consistent in their preferences. Could it be the case that
people are also consistent in the way they perform in reasoning tasks, and can we
predict their answers (including errors) in the aforementioned reasoning domains? We
will explore this by using collaborative filtering to predict participants’ behavior in
reasoning tasks.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>The experimental setting</title>
        <p>For this model, we will use an unpublished data set from an online experiment
conducted at the Cognitive Computation Lab (University of Freiburg). It includes data
from 140 subjects which completed all 64 syllogistic tasks. Each subject was
presented with two premises, and had to choose between nine answer options (the eight
mood/figure combinations, plus the ninth option being No Valid Conclusion).
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>The model</title>
        <p>In our setting, the users are the 140 subjects of the study, and the items are the 64
tasks. We define the similarity function as follows:
(1)
where N represents the amount of questions which were answered by both subjects.
As we can see, similarity is a function between 0 and 1.</p>
        <p>We start by randomly selecting 14 subjects for which there exists at least one other
subject with a similarity of 0.6 or higher, and then randomly deleting 10 of their
answers. These will be the answers which have to be predicted.</p>
        <p>The model computes the pair-wise similarities between subjects, and then
whenever for the current subject there is a missing answer, it identifies all subjects in its
neighborhood (i.e., subjects with a similarity higher than 0.35) which have answered
that task, and performs a “weighted voting” as following:
for answer in possible_answers:
for user in users:
where sim[user] represents the similarity of our subject with the user which we
are currently computing, and given[user] is a binary attribute showing whether</p>
      </sec>
      <sec id="sec-3-4">
        <title>Results</title>
        <p>the user gave this answer to the task or not. We perform this weighting inspired by the
intuition that answers given by more similar subjects should matter more. Then, the
answer with the highest value is the predicted one.</p>
        <p>The model is very simple, and it does not include any learning, its performance is,
however, fairly accurate. It is important to notice that the model predicts one out of
nine possible options, so a model which is simply guessing would be on average
correct in about 11% of times. Our model compares the predicted answers to the true
ones, and reports the percentage of correctly predicted answers.</p>
        <p>In order to interpret the result better it would be useful to compare the performance
of our model to other models or theoretical predictions. As we already stated, most
theories do not focus on individual answer predictions, but on most chosen answers.
For example, a theory can state that for the premises All A are B, Some B are C then
people draw the answers Some A are C, Some C are A or All A are C. We try to see
what these theories would predict for our individual missing answers, and we use the
relaxation that if the missing answer is one of the predicted answers from the theory,
then it is counted as correct. We notice that this is quite a big relaxation, since there
are theories which predict three to four answers for the same pair of premises, which
means they would of course achieve a better accuracy than our model which always
predicts just one answer. We calculate the accuracy of the predictions of theories
based on illicit conversions, verbal models, and mental models, as well as the
predictions of mReasoner, an implementation of the mental models theory.</p>
        <p>One thing to keep in mind is that for some syllogisms there is more than one valid
answer, however subjects could select in our experiment only one answer. This can
cause a difficulty for our comparison as we need to deal with cognitive theories that
often predict up to four or five answers per syllogism. For this reason, we construct
two other versions of our model. Instead of predicting only one answer, we checked
what would be the accuracy of the prediction if we predict the top two and top three
most voted answers. We repeat the procedure 100, 300 and 500 times (to check if
results converge); the results are reported in Table 1:
100 runs
300 runs
500 runs</p>
        <sec id="sec-3-4-1">
          <title>Exact 0.68 0.69 0.69</title>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>Discussion</title>
        <p>The results show that our model which predicts the exact answer does not only
perform reliably better than chance, but even manages to outperform the theoretical
predictions based on illicit conversions, which for almost half of the syllogisms predicts
more than one answer. Furthermore, we notice that our model with the two most
voted predictions outperforms the predictions of the verbal models as well as of
mReasoner, which is right now one of the state-of-the-art predictors for syllogistic
reasoning. Another thing which is important to notice is that our model reaches the same
accuracy even if we delete 32 (out of the 64) answers for up to 50% of the
participants, showing robust performance.</p>
        <p>We notice that the top performance is achieved by the predictions made by the
mental models theory. However, it is important to notice that for almost half of the
syllogisms this theory predicts four or even five answers, which means it has an
advantage for this type of metric. Still, our model which predicts the top 3 answers (still
less than the mental models predictions) achieves the same performance.</p>
        <p>mReasoner is an implementation which is based on mental models, but it has some
parameters which limit the number of predicted answers for each syllogism (it
predicts one answer for 7 syllogisms, and more than two for 16 syllogisms). In this
comparison, we used the default setting for mReasoner, and we see that our model which
predicts the top two answers has a better performance.</p>
        <p>
          Khemlani and Johnson-Laird [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] propose a model where mReasoner learns
parameters for individual subjects in a small dataset consisting of 20 participants, and
then simulates the answers of each subject and compares them to the true answers.
They report a mean correlation to the data of 0.7, which means on average in 70% of
the cases mReasoner made the right prediction. This result is comparable to our basic
model, but built on general cognitive principles. Both approaches differ in their
methodology: Our approach requires participants data to classify and predict other
reasoners and does not have cognitive principles, while on the other hand mReasoners is
built on cognitive principles but trains the system parameters on the whole dataset, so
it is not actually predicting. A combination of both methods to reach a “prediction”
based on cognitive principles is important.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and future steps</title>
      <p>
        These results show that collaborative filtering can help in predicting individual
performance for reasoning tasks, but also that there are new challenges (especially by the
performance boost when considering the top two predictions). First of all, it will be
interesting to test the same model with data from other reasoning domains, e.g., the
Wason selection task [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. This would allow us to test for consistency across different
reasoning domains. Secondly, as we mentioned the model is simple, it would be
interesting to build a more adaptive model which learns from the subjects’ answers and
can identify cognitive principles. This could be achieved by analyzing potential
reasons for differences in performance, combined to using more advanced techniques
from machine learning to build the recommender system.
      </p>
      <p>
        One alternative would be to formalize the tasks by using ternary logic, and then
learn how different subjects map logical operators to truth tables. Ternary logic has
shown to provide high flexibility in modeling Wason’s selection task [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. Another
alternative would be to include theories’ predictions to the task, and check whether a
subject is consistent with the predictions of a certain theory (i.e. we would find
similarities with theoretical predictions rather than with other subjects). This would also
help us for cases where there are not enough subjects to build informative similarity
functions among them.
      </p>
      <p>
        We tried to use machine learning techniques to cluster the data in order to identify
potential reasoning profiles, however the dataset seems to be too diverse. A method
called fcclusterdata, a hierarchal clustering technique from the sckit-learn package
[
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] in Python, identifies more than 40 clusters (for the 140 participants), whereas by
using the k-medoids technique, in which we can specify the number of clusters, for up
to 6 clusters the similarity of subjects in the cluster remains low and we do not
achieve better performance. Studies [
        <xref ref-type="bibr" rid="ref17 ref26">17, 26</xref>
        ] have identified reasons which might lead
to individual differences such as level of intellect, memory capacity etc. Our intuition
is that although these reasons are similar for different individuals, the way they are
presented in people makes it difficult to create clusters. For example, an individual
can have high intellectual capacity but bad memory, another one medium intellectual
capacity and very good memory, and so on. This is why we think that an approach
which focuses on finding similar reasoners for each individual can be more effective.
      </p>
      <p>Reasoners are relatively consistent in their performance in syllogistic reasoning,
since some tend to give similar answers and often predictable mistakes. This means it
is possible to build reasoning models which can identify a person’s reasoning pattern,
and exploit it to better understand the overall reasoning process. This is exactly what
our simple model does, and in its relaxed version it manages to be as good as state of
the art complex reasoning models.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Chapman</surname>
            ,
            <given-names>L. J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chapman</surname>
            ,
            <given-names>J. P.</given-names>
          </string-name>
          <article-title>Atmosphere effect re-examined</article-title>
          .
          <source>Journal of experimental psychology</source>
          ,
          <volume>58</volume>
          (
          <issue>3</issue>
          ),
          <fpage>220</fpage>
          . (
          <year>1959</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Cheng, P. W., &amp;
          <string-name>
            <surname>Holyoak</surname>
            ,
            <given-names>K. J.</given-names>
          </string-name>
          <article-title>Pragmatic reasoning schemas</article-title>
          .
          <source>Cognitive psychology</source>
          ,
          <volume>17</volume>
          (
          <issue>4</issue>
          ),
          <fpage>391</fpage>
          -
          <lpage>416</lpage>
          . (
          <year>1985</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>St. B. T.</surname>
          </string-name>
          <article-title>Heuristic and analytic processes in reasoning</article-title>
          .
          <source>British Journal of Psychology</source>
          ,
          <volume>75</volume>
          ,
          <fpage>451</fpage>
          -
          <lpage>468</lpage>
          . (
          <year>1984</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>St</surname>
            .
            <given-names>B. T.</given-names>
          </string-name>
          <article-title>Bias in human reasoning: Causes and consequences</article-title>
          . Hillsdale, NJ: Erlbaum. (
          <year>1989</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>J. S. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Newstead</surname>
            ,
            <given-names>S. E.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Byrne</surname>
          </string-name>
          , R. M.
          <article-title>Human reasoning: The psychology of deduction</article-title>
          . Psychology Press. (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Gigerenzer</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Hug</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Domain-specific reasoning: Social contracts, cheating, and perspective change</article-title>
          .
          <source>Cognition</source>
          ,
          <volume>43</volume>
          (
          <issue>2</issue>
          ),
          <fpage>127</fpage>
          -
          <lpage>171</lpage>
          . (
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          <article-title>Models of deduction</article-title>
          .
          <source>Reasoning: Representation and process in children and adults</source>
          ,
          <fpage>7</fpage>
          -
          <lpage>54</lpage>
          . (
          <year>1975</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          <article-title>Mental models: Towards a cognitive science of language, inference, and consciousness (No. 6)</article-title>
          . Harvard University Press. (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Steedman</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>The psychology of syllogisms</article-title>
          .
          <source>Cognitive psychology 10</source>
          .1:
          <fpage>64</fpage>
          -
          <lpage>99</lpage>
          . (
          <year>1978</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Wason</surname>
            ,
            <given-names>P. C.</given-names>
          </string-name>
          <article-title>A theoretical analysis of insight into a reasoning task</article-title>
          .
          <source>Cognitive Psychology</source>
          ,
          <volume>1</volume>
          (
          <issue>2</issue>
          ),
          <fpage>134</fpage>
          -
          <lpage>148</lpage>
          . (
          <year>1970</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Kaufman</surname>
          </string-name>
          , L., &amp;
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P. J. Finding</given-names>
          </string-name>
          <article-title>Groups in Data: An Introduction to Cluster Analysis</article-title>
          , Wiley New York Google Scholar. (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Khemlani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          <article-title>Theories of the syllogism: A meta-analysis</article-title>
          .
          <source>Psychological Bulletin</source>
          , Vol
          <volume>138</volume>
          (
          <issue>3</issue>
          ),
          <source>May</source>
          <year>2012</year>
          ,
          <fpage>427</fpage>
          -
          <lpage>457</lpage>
          . (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Khemlani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Johnson-Laird</surname>
            ,
            <given-names>P. N.</given-names>
          </string-name>
          <article-title>How people differ in syllogistic reasoning</article-title>
          .
          <source>In Proceedings of the 36th Annual Conference of the Cognitive Science Society</source>
          . Austin,
          <source>TX: Cognitive Science Society</source>
          . (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Newell</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Simon</surname>
            ,
            <given-names>H. A.</given-names>
          </string-name>
          <article-title>Human problem solving</article-title>
          (Vol.
          <volume>104</volume>
          , No. 9). Englewood Cliffs, NJ: Prentice-Hall.
          <article-title>(</article-title>
          <year>1972</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Oaksford</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chater</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <article-title>A rational analysis of the selection task as optimal data selection</article-title>
          .
          <source>Psychological Review</source>
          ,
          <volume>101</volume>
          ,
          <fpage>608</fpage>
          -
          <lpage>631</lpage>
          . (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Oaksford</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chater</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <article-title>Theories of reasoning and the computational explanation of everyday inference</article-title>
          .
          <source>Thinking &amp; Reasoning</source>
          ,
          <volume>1</volume>
          (
          <issue>2</issue>
          ),
          <fpage>121</fpage>
          -
          <lpage>152</lpage>
          . (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Oberauer</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Süß</surname>
            ,
            <given-names>H. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilhelm</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Sander</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <article-title>Individual differences in working memory capacity and reasoning ability</article-title>
          . Variation in working memory,
          <fpage>49</fpage>
          -
          <lpage>75</lpage>
          . (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Pachur</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bröder</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Marewski</surname>
            ,
            <given-names>J. N.</given-names>
          </string-name>
          <article-title>The recognition heuristic in memory‐based inference: is recognition a non‐compensatory cue</article-title>
          ?
          <source>Journal of Behavioral Decision Making</source>
          ,
          <volume>21</volume>
          (
          <issue>2</issue>
          ),
          <fpage>183</fpage>
          -
          <lpage>210</lpage>
          . (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          , ... &amp;
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>12</volume>
          (Oct),
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          . (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Polk</surname>
            ,
            <given-names>T. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Newell</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <article-title>Deduction as verbal reasoning</article-title>
          .
          <source>Psychological Review</source>
          ,
          <volume>102</volume>
          (
          <issue>3</issue>
          ),
          <fpage>533</fpage>
          . (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Ragni</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dietz</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kola</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Hölldobler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <article-title>Two-Valued Logic is Not Sufficient to Model Human Reasoning, but Three-Valued Logic is: A Formal Analysis</article-title>
          .
          <source>Bridging 2016 - Bridging the Gap between Human and Automated Reasoning</source>
          ,
          <volume>1651</volume>
          :
          <fpage>61</fpage>
          -
          <lpage>73</lpage>
          . (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Revlis</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>Two models of syllogistic reasoning: Feature selection and conversion</article-title>
          .
          <source>Journal of Verbal Learning and Verbal Behavior</source>
          ,
          <volume>14</volume>
          (
          <issue>2</issue>
          ),
          <fpage>180</fpage>
          -
          <lpage>195</lpage>
          . (
          <year>1975</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Resnick</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Varian</surname>
            ,
            <given-names>H. R.</given-names>
          </string-name>
          <article-title>Recommender systems</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>40</volume>
          (
          <issue>3</issue>
          ),
          <fpage>56</fpage>
          -
          <lpage>58</lpage>
          . (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Rips</surname>
            ,
            <given-names>L. J.</given-names>
          </string-name>
          <article-title>Cognitive processes in propositional reasoning</article-title>
          .
          <source>Psychological review</source>
          ,
          <volume>90</volume>
          (
          <issue>1</issue>
          ),
          <fpage>38</fpage>
          . (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Rips</surname>
            ,
            <given-names>L. J.</given-names>
          </string-name>
          <article-title>The psychology of proof: Deductive reasoning in human thinking</article-title>
          . MIT Press.
          <article-title>(</article-title>
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Stanovich</surname>
            ,
            <given-names>K. E.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>West</surname>
            ,
            <given-names>R. F.</given-names>
          </string-name>
          <article-title>Individual differences in rational thought</article-title>
          .
          <source>Journal of experimental psychology: general</source>
          ,
          <volume>127</volume>
          (
          <issue>2</issue>
          ),
          <fpage>161</fpage>
          . (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Tversky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Kahneman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Availability: A heuristic for judging frequency and probability</article-title>
          .
          <source>Cognitive psychology</source>
          ,
          <volume>5</volume>
          (
          <issue>2</issue>
          ),
          <fpage>207</fpage>
          -
          <lpage>232</lpage>
          . (
          <year>1973</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Wason</surname>
            ,
            <given-names>P. C.</given-names>
          </string-name>
          <string-name>
            <surname>Reasoning</surname>
          </string-name>
          . New Horizons in Psychology. pp.
          <fpage>135</fpage>
          -
          <lpage>151</lpage>
          . (
          <year>1966</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>