<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>C. Beyer);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Challenges for Active Feature Acquisition and Imputation on Data Streams</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Christian Beyer</string-name>
          <email>christian.beyer@ovgu.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maik Büttner</string-name>
          <email>maik.buettner@ovgu.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Myra Spiliopoulou</string-name>
          <email>myra@ovgu.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Otto-von-Guericke-University Magdeburg</institution>
          ,
          <addr-line>Universitätsplatz 2, 39106 Magdeburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Two popular methods for dealing with missing feature values are active feature acquisition as well as imputation. Both methods often require an understanding of a feature's relationship to the target variable as well as to all the other features. Developing such an understanding is time-consuming and challenging in a static setting, but becomes much more complicated in a data stream scenario. Additional challenges are concept drift, feature drift, incorporating feature costs, dealing with complex types of missingness, and the need for imputation models that can be updated eficiently. In this work, we will discuss these challenges as well as challenges that appear downstream when devising stream-applicable solutions. The goal is to provide a current overview and inspire discussion as well as further research in this field.</p>
      </abstract>
      <kwd-group>
        <kwd>active learning</kwd>
        <kwd>data streams</kwd>
        <kwd>imputation</kwd>
        <kwd>active feature acquisition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Many machine learning models can only be trained and create predictions if all the features of
an instance are available. This imposes a need for methods to replace missing feature values
during training and testing. Imputation is the most popular approach to deal with missing
values, where the missing feature values are estimated using heuristics and models, that either
rely on a feature’s distribution or its relationship to other features [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Another approach to deal
with missing values is Active Feature Acquisition (AFA), where the real feature values can be
purchased from a costly oracle under budget constraints [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. An oracle could be a costly subject
matter expert that has to be inquired or a lab test that has to be done. Both approaches have
diferent advantages and disadvantages, see Table
1. This makes them applicable in diferent
scenarios and often it would make sense to use a mix of both methods. For example, if an
instance has a feature missing, that is strongly correlated with another feature that is available,
then using imputation might be a good strategy. On the other hand, if the missing feature
cannot be predicted well by available features, it might require a costly lab test or the opinion
of a subject matter expert to determine the real feature value. This hypothetical scenario shows,
that knowledge about the features and their relationship to one another and their relationship
to the target variable is needed. The inter-feature relationships are needed in order to build
proper models for imputation as well as to guide AFA methods toward purchases of features
CEUR
that cannot be inferred from available ones. The relationship to the target variable is needed
to know if a certain feature is even relevant to the task at hand, which is often called the  
of a feature. If it is not relevant then we can safely remove it, as the feature value should not
influence the final decision. In case it is relevant, then we should consider purchasing it if it
cannot be imputed with high confidence. In contrast to a static setting, in a data stream all these
relationships might be subject to change and imputation models that map the inter-feature
relationship need to be updated often and in a timely manner. These and subsequent challenges
will now be discussed in more detail.
      </p>
      <sec id="sec-1-1">
        <title>Approach</title>
        <sec id="sec-1-1-1">
          <title>Imputation</title>
        </sec>
        <sec id="sec-1-1-2">
          <title>Active Feature Acquisition</title>
        </sec>
      </sec>
      <sec id="sec-1-2">
        <title>Advantages Disadvantages</title>
        <p>- biased estimates
---thfcneaaosnwtcuohcossotumleapldlayartebadestaeoptpAlFieAd to ----daircMmetaaqonpuspubtirrteomaepwsteieortrrehontpnoiemrgdessesetohnnotladytidhveaessidgtaonteamdaftocrh
MCAR
- costly
- real values - can be slow
- no need for representative data - can usually only be applied to a
fraction of the data set</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. The Challenge of Dealing with Diferent Types of</title>
    </sec>
    <sec id="sec-3">
      <title>Missingness</title>
      <p>
        Almost all stream-related publications deal with one type of missingness which is missing
completely at random (MCAR) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This makes other types of missingness an under-researched
area. MCAR means that the missingness of a certain feature is independent of factors within
the data set and cannot be explained by outside factors as well. Though easy to model, it is also
the least likely case to be encountered, as in most cases the missingness either depends on the
variable itself or on variables inside or outside the data set. An example of the former would be,
if older people were less likely to state their age, and an example of the latter would be if people
of a certain gender would tend to skip certain questions. The challenges are to detect which
type of missingness we are confronted with and develop stream-applicable methods that can
handle missingness apart from MCAR. In [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] the authors considered a special scenario where all
the features of an instance are missing and are purchased iteratively and while their approach
is designed for streams, it seems very inadequate to deal with drift among the features. Another
interesting solution is presented in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], where the authors propose a deep ladder imputation
network that can handle any kind of missing data and also deal with high degrees of missingness.
Unfortunately, it is again ill-suited for streaming data containing drift. Real-time induction and
updating of the model are additional open challenges.
      </p>
    </sec>
    <sec id="sec-4">
      <title>3. The Challenge of Dealing with Drift</title>
      <p>
        In an incomplete stream two types of drift may occur: feature drift and concept drift. Feature drift
occurs if the distribution of a feature changes or if the relationship of the feature to the target
variable changes [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Feature drift can occur abruptly, gradually, or shifting and necessitates an
update of the imputation models as well as the   estimates used by AFA. In order to address
potential feature drift, we first have to detect it. This constitutes another challenge in itself as
drift detection algorithms are specialized in detecting certain types of drift and there is no free
lunch [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Once feature drift has been detected it is important to forget outdated information
and to update the imputation models and   estimates. There are also simple solutions like
windowing techniques [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] but windows of static length have the disadvantages that we might
miss moments of feature drift and apply outdated models for a while or that the imputation
and   estimates are subpar because the available training data is artificially restricted by the
window length.
      </p>
      <p>
        A temporal change in the distribution of the target variable or a change in the relationship of
the target variable to other features is called concept drift[
        <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
        ]. Concept drift does not afect
imputation directly, as the target variable is usually not used, but a change in the target variables
distribution could exacerbate the problem of biases in the imputed values. For example, if we
consider an imputation method that always replaces a missing feature with the feature mean
and a highly skewed data set, where the majority class often has values around the feature mean
associated with it, then we will produce only a few prediction errors. If concept drift happens so
that the minority class is more prominent around that feature’s mean then we might introduce
a lot of errors, because the predictions will now favor the minority class. The problems for AFA
are more obvious as a feature’s   is supposed to inform us how valuable it is for solving the
task at hand, which often means how well it helps in separating the classes [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. If classes now
start to overlap or change abruptly, then   estimates will need to be updated and changed
accordingly. This again necessitates first of all that we recognize when drift happens.
      </p>
      <p>
        Deep-Learning-based approaches are becoming increasingly popular as they achieve high
performances [
        <xref ref-type="bibr" rid="ref10 ref3 ref4">3, 4, 10</xref>
        ] but are highly susceptible to the issue of drift. Their need for a lot of
representative data and computational time to adapt to new concepts makes them ill-suited for
such scenarios.
      </p>
    </sec>
    <sec id="sec-5">
      <title>4. The Challenge to Induce Imputation Models in Near Real</title>
    </sec>
    <sec id="sec-6">
      <title>Time</title>
      <p>
        If we consider the detection of feature and concept drift solved then we are still left with the
need for imputation models that can be updated in almost real-time. This excludes or makes
the application of several prominent imputation methods like MICE [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and methods based
on deep neural networks [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] much harder because they require multiple runs over the same
training data which can be very time-consuming. The structure of inter-feature relationships
could be modeled with a Bayesian network [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and used for imputation but online versions
have shown to be subpar to static versions [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and it is also challenging to adapt Bayesian
Networks to diferent types of drift [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], especially shifting drift. One proposed solution to
make algorithms designed for static data viable on data streams, is the usage of windowing
techniques [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to restrict training data. However, these cannot always be applied especially
when we want to employ deep learning-based methods which promise high imputation quality.
Knowledge about the inter-feature relationships would also be of high value in an AFA setting.
One drawback of stream applicable   estimates [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is, that these estimates are independent
for each feature and therefore ignore all feature-to-feature relationships which could potentially
be exploited. It could therefore happen that we inquire a costly oracle to provide a feature value
that could have been predicted very well by other features that were available.
      </p>
    </sec>
    <sec id="sec-7">
      <title>5. The Challenge of Dealing with Feature Costs</title>
      <p>
        Features can have varying costs, for example, measuring a patient’s temperature requires less
costly materials and less skill than running an MRI. Costs do not have to be monetary. They
can also describe the time or expertise required to acquire a feature. These varying costs
might introduce an additional bias towards cheaper features in the selection process of feature
acquisitions when the budget is a further constraint to consider. Such biases worsen the problem
of the trade-of between exploration and exploitation whereby neglecting to purchase specific
features due to their inhibiting cost might delay the detection of new feature concepts and
inter-feature relationships. However, incorporating feature costs has been shown to improve
predictive performance in static settings under budget constraints [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Our early work on
data streams supports this notion [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. We pointed out that it requires more complex  
functions. These should take the inter-feature relationships into account. Feature costs are also
a motivation to tackle the challenge of combining AFA and imputation in an intelligent manner
so that the budget is only spent to purchase feature values that cannot be imputed well. In the
case of missing labels, queries for both labels and features may be combined to allow learning
agents themselves to decide on the trade-of of prioritizing training imputation models, training
prediction models, adapting models to drifts, and saving budget.
      </p>
    </sec>
    <sec id="sec-8">
      <title>6. Conclusion</title>
      <p>In this short work, we motivate challenges that impede the application of imputation and AFA
methods on data streams, especially when we want to apply them in a joint framework. The
main challenges are:
• Need for imputation models that can handle any type of missingness
• Need for fast, high-performance imputation models that can be updated incrementally
• Need for a general feature and concept drift detection
• Modelling the inter-feature relationship in the face of diferent kinds of drift
• Need for AFA methods that take feature costs into account</p>
      <p>We hope this work will encourage discussion, as well as future research that addresses these
challenges.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>W.-C.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-F.</given-names>
            <surname>Tsai</surname>
          </string-name>
          ,
          <article-title>Missing value imputation: a review and analysis of the literature (</article-title>
          <year>2006</year>
          -2017),
          <source>Artificial Intelligence Review</source>
          <volume>53</volume>
          (
          <year>2020</year>
          )
          <fpage>1487</fpage>
          -
          <lpage>1509</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B.</given-names>
            <surname>Settles</surname>
          </string-name>
          ,
          <article-title>Active learning literature survey</article-title>
          ,
          <source>Technical Report 1648</source>
          , University of WisconsinMadison Department of Computer Sciences,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kachuee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Goldstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kärkkäinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Darabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sarrafzadeh</surname>
          </string-name>
          ,
          <article-title>Opportunistic learning: Budgeted cost-sensitive learning from data streams</article-title>
          ,
          <source>in: 7th International Conference on Learning Representations, ICLR</source>
          <year>2019</year>
          ,
          <article-title>New Orleans</article-title>
          , LA, USA, May 6-
          <issue>9</issue>
          ,
          <year>2019</year>
          , OpenReview.net,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Hallaji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Razavi-Far</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saif</surname>
          </string-name>
          , Dlin:
          <article-title>Deep ladder imputation network</article-title>
          ,
          <source>IEEE Transactions on Cybernetics</source>
          <volume>52</volume>
          (
          <year>2021</year>
          )
          <fpage>8629</fpage>
          -
          <lpage>8641</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Barddal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. M.</given-names>
            <surname>Gomes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Enembreck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pfahringer</surname>
          </string-name>
          ,
          <article-title>A survey on feature drift adaptation: Definition, benchmark, challenges and future directions</article-title>
          ,
          <source>Journal of Systems and Software</source>
          <volume>127</volume>
          (
          <year>2017</year>
          )
          <fpage>278</fpage>
          -
          <lpage>294</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kantardzic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Sethi</surname>
          </string-name>
          ,
          <article-title>No free lunch theorem for concept drift detection in streaming data classification: A review</article-title>
          ,
          <source>Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</source>
          <volume>10</volume>
          (
          <year>2020</year>
          )
          <article-title>e1327</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>An exploration of online missing value imputation in nonstationary data stream</article-title>
          ,
          <source>SN Computer Science</source>
          <volume>2</volume>
          (
          <year>2021</year>
          ).
          <source>doi:1 0 . 1 0 0 7 / s 4 2</source>
          <volume>9 7 9 - 0 2 1 - 0 0 4 5 9 - 1</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gama</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Zhang, Learning under concept drift: A review, IEEE Transactions on Knowledge and Data Engineering (</article-title>
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pfahringer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Barddal</surname>
          </string-name>
          ,
          <article-title>Addressing feature drift in data streams using iterative subset selection</article-title>
          ,
          <source>ACM SIGAPP Applied Computing Review</source>
          <volume>19</volume>
          (
          <year>2019</year>
          )
          <fpage>20</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kossen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cangea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Vértes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jaegle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Patraucean</surname>
          </string-name>
          , I. Ktena,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tomasev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Belgrave</surname>
          </string-name>
          ,
          <article-title>Active acquisition for multimodal temporal data: A challenging decision-making task</article-title>
          ,
          <source>Transactions on Machine Learning Research</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Van Buuren</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Groothuis-Oudshoorn, mice: Multivariate imputation by chained equations in r</article-title>
          ,
          <source>Journal of statistical software 45</source>
          (
          <year>2011</year>
          )
          <fpage>1</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Howey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. D.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Naamane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. N.</given-names>
            <surname>Reynard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Pratt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Cordell</surname>
          </string-name>
          ,
          <article-title>A bayesian network approach incorporating imputation of missing data enables exploratory analysis of complex causal biological relationships</article-title>
          ,
          <source>PLoS Genetics</source>
          <volume>17</volume>
          (
          <year>2021</year>
          )
          <article-title>e1009811</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ratnapinda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Druzdzel</surname>
          </string-name>
          ,
          <article-title>Learning discrete bayesian network parameters from continuous data streams: What is the best strategy?</article-title>
          ,
          <source>Journal of Applied Logic</source>
          <volume>13</volume>
          (
          <year>2015</year>
          )
          <fpage>628</fpage>
          -
          <lpage>642</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Liu,
          <article-title>Learning non-stationary dynamic bayesian network structure from data stream</article-title>
          ,
          <source>in: 2019 IEEE Fourth International Conference on Data Science in Cyberspace (DSC)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>128</fpage>
          -
          <lpage>134</lpage>
          .
          <source>doi:1 0 . 1 1 0 9 / D S C . 2 0</source>
          <volume>1 9 . 0 0 0 2 7 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>O.</given-names>
            <surname>Kaminska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Klonecki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kaczmarek-Majer</surname>
          </string-name>
          ,
          <article-title>Feature selection in bipolar disorder episode classification using cost-constrained methods</article-title>
          ,
          <source>in: Artificial Intelligence in Medicine</source>
          , Springer International Publishing,
          <year>2023</year>
          . To be published.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Büttner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Beyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Spiliopoulou</surname>
          </string-name>
          ,
          <article-title>Reducing missingness in a stream through costaware active feature acquisition</article-title>
          ,
          <source>in: 2022 IEEE 9th International Conference on Data Science and Advanced Analytics (DSAA)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>