<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Relational Arti cial Intelligence</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Molham Aref</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>RelationalAI</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>USA molham.aref@gmail.com</string-name>
        </contrib>
      </contrib-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In this talk, I will make the case for a rst-principles approach to machine
learning over relational databases that exploits recent development in database
systems and theory. The input to learning classi cation and regression models
is de ned by feature extraction queries over relational databases. We cast the
machine learning problem as a database problem by decomposing the learning
task into a batch of aggregates over the feature extraction query and by
computing this batch over the input database. The performance of this approach
bene ts tremendously from structural properties of the relational data and of
the feature extraction query; such properties may be algebraic (semi-ring),
combinatorial (hypertree width), or statistical (sampling). This translates to several
orders-of-magnitude speed-up over state-of-the-art systems. This work is based
on collaboration with Hung Q. Ngo (RelationalAI), Mahmoud Abo-Khamis
(RelationalAI), Ryan Curtin (RelationalAI), Dan Olteanu (Oxford), Maximilian
Schleich (Oxford), Ben Moseley (CMU), and XuanLong Nguyen (Michigan) and
other members of the RelationalAI team and faculty network.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>