<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Method Selection for Feature-Poor Bipartite Graphs: From Traditional Baselines to Graph Neural Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pavel Prochazka</string-name>
          <email>P@0.50</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniil Buchko</string-name>
          <email>dabuchko@cisco.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Practical Guidelines, Bipartite Graphs, Graph Representation Learning, Node Classification, Empirical Analysis</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cisco Systems, Inc.</institution>
          ,
          <addr-line>Prague</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Bipartite graphs naturally arise in structured data applications such as recommendation systems and knowledge graphs. While Graph Neural Networks (GNNs) excel when rich node features are available, many real-world scenarios lack such features, leaving only structural information. This raises a fundamental practical question: How should practitioners approach method selection for learning on feature-poor bipartite graphs? We develop a systematic framework grounded in five empirical hypotheses about method efectiveness, validated through comprehensive comparison of 16 methods across six real-world datasets. Our results reveal that method efectiveness is highly dataset-dependent, with no single approach dominating all scenarios. Traditional methods like Label Propagation provide excellent starting points due to their computational eficiency (2-3 orders of magnitude faster), while well-established GNNs like HGCN ofer competitive performance when additional resources are available. Our framework provides systematic guidance through four tiers of increasing complexity, with clear decision points based on confident prediction assessment and resource allocation principles that address real-world deployment constraints.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Graph Neural Networks (GNNs) have demonstrated remarkable success for learning on graph-structured
data when rich node features are available. However, many real-world bipartite graph
scenarios—including recommendation systems, citation networks, and knowledge graphs—naturally lack meaningful
node features, leaving only structural information for learning. This characteristic fundamentally
challenges the conventional assumption that GNNs consistently outperform traditional graph learning
methods, since their primary advantage of joint structure-feature learning is diminished. Moreover,
GNNs introduce practical limitations that become pronounced in feature-poor scenarios: extensive
computational requirements (2–3 orders of magnitude higher than traditional methods), complex
hyperparameter optimization, and implementation challenges that may not be justified when their core
advantage is reduced. This raises the fundamental question: How should practitioners systematically
approach method selection for learning on feature-poor bipartite graphs?</p>
      <p>We address this challenge by developing a systematic framework grounded in five empirical
hypotheses about method efectiveness in feature-poor scenarios. Our comprehensive evaluation compares 16
methods across six real-world datasets, revealing that method efectiveness is highly dataset-dependent
with no universal winner across scenarios. Traditional methods like Label Propagation provide
excellent starting points due to their computational eficiency and reliable optimization, while established
GNNs ofer competitive performance when additional resources are justified. Our key contribution is a
four-tier progression framework with clear decision points based on confident prediction assessment,
enabling practitioners to balance performance requirements against computational constraints and
implementation complexity in real-world deployment scenarios.</p>
      <p>CEUR
Workshop</p>
      <p>ISSN1613-0073</p>
      <p>Paper Organization:</p>
      <p>We formalize the bipartite graph learning problem, then present our
hypothesisdriven framework for systematic method selection. Related work provides tier-based method
categorization and dataset context, followed by empirical evidence supporting our framework hypotheses and
conclusions with future research directions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem Formulation</title>
      <sec id="sec-2-1">
        <title>2.1. Notation and Definitions</title>
        <p>We consider bipartite graphs  = ( ,  , )
where  and  are disjoint node sets and  ⊆  × 
represents
edges. The bipartite structure is characterized by the incidence matrix  ∈ {0, 1} | |×| | , where   = 1
if there exists an edge between   ∈  and   ∈  . No additional node features are provided beyond
structural information encoded in  .</p>
        <p>Methods requiring unipartite graph input (such as GraphSAGE or label propagation) convert the
bipartite graph to a unipartite representation using the adjacency matrix
 =
[ 
0 
0</p>
        <p>For node   ∈  , we denote its neighborhood in  as  (

) = {  ∈  ∶   = 1}. Feature-based methods
ifrst compute dense node representations through techniques such as matrix factorization or spectral
embedding of  . Computation of dense representation introduces additional computational overhead
compared to methods. These methods operate directly on the sparse structural information and they
use one-hot feature matrix ( ) representation.</p>
        <p>We examine two fundamental learning tasks on these feature-poor bipartite graphs:
Classification Task:</p>
        <sec id="sec-2-1-1">
          <title>Given a training set  ⊂  with known labels, we predict labels for nodes in  .</title>
          <p>Performance is measured using classification accuracy, which assesses average performance across all
predictions.</p>
          <p>Retrieval Task: For datasets with multi-class labels, we formulate retrieval as a series of
oneversus-rest binary classification problems. For each class  ∈ {1, 2, … , } , given a training set   + ⊂ 
containing only positive examples for class  , methods must retrieve the top- nodes most likely to
where models are most confident. The final retrieval performance is computed as the average over all</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Problem Statement</title>
        <p>tional overhead and greater reliability
outperform traditional approaches
The central research question addressed in this paper is: How should practitioners systematically select
methods for learning on feature-poor bipartite graphs, where Graph Neural Networks lose their primary
advantage of joint structure-feature learning?</p>
        <p>This question becomes particularly relevant because:
• GNNs typically require dense feature representations, making them computationally expensive
compared to methods that operate directly on sparse structural information
• Traditional graph methods may ofer competitive performance with significantly lower
computa• The absence of rich node features challenges the conventional assumption that GNNs consistently
Rather than focusing on developing new algorithms or providing theoretical analysis of method
capabilities, this paper addresses the practical challenge of eficient method selection
in
resourceconstrained settings. Our goal is to provide systematic, evidence-based guidance that helps practitioners
navigate the trade-ofs between computational cost, implementation complexity, and performance
across diferent method categories.</p>
        <p>Scope and Contribution: This work focuses specifically on providing practical method selection
guidance through systematic empirical analysis. We aim to share actionable insights that inform
decision-making when computational budgets, implementation timelines, and optimization resources
are limited—constraints that are common in real-world deployment scenarios but often overlooked in
academic evaluations focused primarily on performance maximization.
3. Method Selection Framework for Feature-Poor Bipartite Graphs
Based on our comprehensive empirical analysis across six bipartite graph datasets, we present a
systematic framework for method selection that addresses the core challenge practitioners face: eficiently
navigating the trade-ofs between computational cost, implementation complexity, and performance
when Graph Neural Networks lose their primary advantage of joint structure-feature learning.</p>
      </sec>
      <sec id="sec-2-3">
        <title>3.1. Empirical Hypotheses</title>
        <p>Our framework is grounded in five key hypotheses derived from systematic evaluation across diverse
bipartite graph scenarios. These hypotheses form the theoretical foundation for our practical guidance,
with confidence levels reflecting the strength of supporting evidence from our empirical study.</p>
        <p>Hypothesis 1: No Universal Winner. Method efectiveness is fundamentally dataset-dependent in
feature-poor bipartite graphs, with no single approach consistently dominating across all scenarios.
This hypothesis provides the core motivation for systematic evaluation rather than assuming GNN
superiority across all bipartite graph learning tasks.</p>
        <p>Hypothesis 2: Implementation Maturity Efects . Implementation reliability and optimization stability
often outweigh theoretical performance advantages, particularly for specialized methods lacking
standard library support. Methods requiring extensive custom implementation frequently exhibit reliability
issues that limit practical applicability.</p>
        <p>Hypothesis 3: Dense Feature Extraction Trade-of . Dense feature extraction through matrix
factorization or spectral embedding could provide performance benefits but at 2–3 orders of magnitude
computational cost compared to methods operating directly on sparse structural information. This
trade-of defines a natural complexity boundary in method selection.</p>
        <p>Hypothesis 4: Simple Method Failure Principle. If simple methods fail to achieve reliable performance
on confident predictions, complex methods rarely provide significant improvements. This principle
guides the critical decision between method progression versus data quality improvement.</p>
        <p>Hypothesis 5: Optimization Budget Constraints. With realistic hyperparameter optimization
budgets, systematic progression through increasing complexity tiers typically provides better return on
investment than directly applying the most sophisticated methods. Simpler methods benefit from more
thorough optimization within fixed time constraints.</p>
      </sec>
      <sec id="sec-2-4">
        <title>3.2. Four-Tier Method Selection Framework</title>
        <p>Our framework organizes methods into four tiers of increasing implementation complexity and
computational requirements. Each tier boundary is motivated by specific empirical hypotheses:</p>
        <p>
          Tier 1 - Structural Baselines: Methods operating directly on bipartite graph structure using one-hot
representations (CSP [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], Label Propagation, One-Hot Logistic Regression). Motivated by Hypothesis 3:
These approaches avoid the computational overhead of dense feature extraction while providing rapid
feasibility assessment with straightforward hyperparameter optimization.
        </p>
        <p>
          Tier 2 - Traditional ML Methods: Traditional machine learning methods requiring dense feature
extraction (Feature-based Logistic Regression, Random Forest, MLP) and simple single-layer neural
networks (HGCN [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]). Motivated by Hypotheses 3 and 5: These methods accept the computational
cost of feature preprocessing but maintain manageable configuration spaces that benefit from thorough
optimization within realistic time constraints.
        </p>
        <p>
          Tier 3 - Established Graph Neural Networks: Well-established GNN architectures with large
configuration spaces (e.g. HGCN with multiple layers [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], HyperND [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], etc.). Motivated by Hypotheses 2 and 5:
These methods ofer sophisticated graph-aware learning through reliable implementations, but their
extensive hyperparameter spaces require significant optimization resources that may limit practical
advantages.
        </p>
        <p>
          Tier 4 - Specialized Experimental Methods: Methods specifically designed for feature-poor scenarios
but lacking mature implementations (VilLain [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], Subgradient methods [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]). Motivated by Hypothesis
2: These approaches may ofer theoretical advantages but present implementation challenges and
reliability concerns that limit practical deployment.
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>3.3. Systematic Selection Strategy</title>
        <p>Our framework provides clear decision points based on the empirical hypotheses:</p>
        <p>Phase 1 - Initial Assessment: Identify whether Tier 1 methods can handle ”simple examples”—test
instances that are very similar to training examples according to the given structural representation.
If they demonstrate this fundamental capability, proceed to Phase 2 for systematic enhancement. If
not, focus on data quality improvement by enhancing the representation (either through improved
graph construction or additional features). This assessment can be operationalized through suficient
retrieval performance on confident predictions (start with  @100 ≤ 0.5 for identification of insuficient
performance, but adapt this threshold to your specific problem characteristics). This principle reflects
Hypothesis 4: when simple methods fail on their most confident predictions, complex architectures
rarely overcome fundamental task limitations.</p>
        <p>Phase 2 - Progressive Enhancement: Based on Hypothesis 5: When Tier 1 methods demonstrate
fundamental capability but insuficient performance for application requirements, progress systematically
through higher tiers. Each tier transition should be justified by clear performance requirements that
outweigh increased computational costs and optimization complexity.</p>
        <p>Resource Allocation Guidance: Based on Hypotheses 2 and 5. Allocate optimization efort
proportionally to method reliability - invest heavily in Tier 1-2 hyperparameter exploration, moderately in Tier 3
given large configuration spaces, and cautiously in Tier 4 given implementation uncertainties.</p>
      </sec>
      <sec id="sec-2-6">
        <title>3.4. Framework Validation and Limitations</title>
        <p>Our empirical evaluation demonstrates that this systematic approach efectively captures
performancecomplexity trade-ofs across diverse bipartite graph scenarios. The framework successfully identifies
when simple methods sufice (high confident prediction performance) versus when sophistication is
warranted (clear progression benefits).</p>
        <p>Confidence Assessment : Hypotheses 1-3 receive strong support across all evaluated datasets, while
Hypotheses 4-5 show consistent but more limited evidence requiring broader validation. The framework
principles appear robust, but specific thresholds and decision points may require adaptation for diferent
domains or task characteristics.</p>
        <p>Practical Impact: When the framework’s assumptions hold, it enables eficient resource allocation and
reduces the risk of over-engineering solutions for tasks where simple methods sufice. When
assumptions fail, systematic evaluation through the tier structure still provides valuable comparative insights
while avoiding premature commitment to complex approaches. If medium-confidence hypotheses (4-5)
prove invalid in specific scenarios - for example, if complex methods succeed when simple methods
fail on confident predictions, or if extensive optimization consistently improves complex method
performance - practitioners should adapt resource allocation and decision thresholds accordingly while
maintaining the systematic tier-based evaluation structure.</p>
        <p>This framework transforms our empirical findings into actionable guidance while acknowledging
the need for continued validation across broader domains and graph types. The systematic progression
strategy addresses real-world constraints of limited computational budgets and optimization time while
providing clear decision points for when to invest in increased method sophistication.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Related Work</title>
      <p>Having established our problem formulation and framework, we position our work within the broader
landscape of bipartite graph learning methods and systematic approach selection for machine learning
tasks.</p>
      <sec id="sec-3-1">
        <title>4.1. Evolution of Bipartite Graph Learning Methods</title>
        <p>The development of bipartite graph learning methods follows a natural progression of increasing
sophistication that aligns with our tier-based framework, as summarized in Table 1.</p>
        <p>
          Tier 1 - Structural Methods: Early collaborative filtering approaches [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] operated directly on bipartite
user-item structures without requiring dense feature representations. Classical label propagation [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]
extended semi-supervised learning to graph structures through iterative message passing on converted
unipartite representations. CSP [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] recently introduced a parameter-free hypergraph variant that
works natively on bipartite structures, implementable with simple SQL operations while avoiding
hyperparameter optimization entirely.
        </p>
        <p>
          Tier 2 - Traditional ML Methods: Traditional machine learning methods leveraging dense
representations emerged with matrix factorization techniques [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and spectral methods for collaborative
ifltering. These approaches accept the computational overhead of feature extraction while maintaining
manageable configuration spaces. Simple neural baselines like MLPs provide structure-agnostic learning
on dense features, while single-layer graph neural networks like HGCN [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] ofer limited architectural
complexity with small configuration spaces.
        </p>
        <p>
          Tier 3 - Established Graph Neural Networks: The rise of Graph Neural Networks marked a paradigm
shift in graph-based learning. Dai et al. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] demonstrated that GNNs can be viewed as learned message
passing algorithms, bridging classical approaches with modern deep learning such as GraphSAGE [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
Extensions to hypergraph and bipartite scenarios include HGCN [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], UniGCNII [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], HyperND [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ],
EHGNN [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], HCHA [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], and AllSetTransformer [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. While these methods demonstrate sophisticated
architectural designs, they require extensive hyperparameter tuning and significant computational
resources.
        </p>
        <p>
          Tier 4 - Specialized Experimental Methods: Recent approaches like VilLain [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and subgradient
hypergraph classifiers [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] specifically target feature-poor scenarios through novel optimization procedures.
However, these methods lack mature implementations in standard libraries, presenting deployment
challenges and reliability concerns that limit practical applicability.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>4.2. Method Selection and Systematic Evaluation Approaches</title>
        <p>Our work contributes to the broader challenge of systematic method selection in machine learning,
particularly for graph-structured data where method choice significantly impacts both performance
and computational requirements.</p>
        <p>
          AutoML and Method Selection: The automated machine learning community has developed systematic
approaches for algorithm selection [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], typically focusing on tabular data with extensive feature
engineering. However, these approaches rarely address the unique challenges of graph-structured data
where structural representation choices fundamentally afect method applicability.
        </p>
        <p>
          Graph Learning Method Comparisons: Prior comparative studies in graph learning have primarily
focused on rich-feature scenarios where GNNs demonstrate clear advantages [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Systematic
evaluation of traditional versus modern methods specifically for feature-poor bipartite graphs remains
underexplored, with most studies assuming GNN superiority rather than providing decision frameworks
for method selection.
        </p>
        <p>
          Practical Guidelines in Machine Learning: The machine learning community increasingly recognizes
the need for practical guidance that considers implementation constraints alongside theoretical
performance [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. Our tier-based framework contributes to this direction by providing systematic decision
points based on computational budgets, optimization reliability, and implementation maturity rather
than focusing solely on performance maximization.
        </p>
        <p>
          Empirical Methodology for Graph Learning: Recent work has begun questioning universal GNN
superiority, particularly when their primary advantage of joint structure-feature learning is diminished [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
Our systematic empirical approach builds on this trend by providing evidence-based method selection
guidance specifically for feature-poor bipartite scenarios.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>4.3. Bipartite Graph Datasets and Evaluation Domains</title>
        <p>Our empirical evaluation spans six real-world bipartite graph datasets that naturally lack rich node
features, requiring methods to learn exclusively from structural information. These datasets represent
two primary domains where feature-poor bipartite graphs commonly arise in practice.</p>
        <p>
          Academic Citation Networks: Five datasets derive from scholarly publication databases, each ofering
diferent structural perspectives on academic relationships. Cora provides dual bipartite representations
connecting papers to authors (Cora-CA) and papers to citation contexts (Cora-CC), demonstrating
how the same underlying data can yield fundamentally diferent bipartite structures [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. CiteSeer and
PubMed follow citation-based approaches in computer science and biomedical domains respectively [
          <xref ref-type="bibr" rid="ref20 ref21">20,
21</xref>
          ], while DBLP represents author-paper collaborations across computer science publications [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
These datasets exhibit substantial structural diversity, with node counts ranging from 2,708 to 41,302
and varying degrees of connectivity.
        </p>
        <p>E-commerce Domain: The Walmart dataset represents product co-purchase relationships, where
bipartite edges indicate items frequently bought together. This domain introduces diferent structural
properties compared to citation networks, with higher edge density (460,629 edges across 88,860
products) and distinct degree distributions that stress-test method generalization across application
domains.</p>
        <p>Dataset Characteristics and Challenges: Our collection exhibits natural diversity in structural
complexity, from sparse citation networks to dense co-purchase graphs. The presence of isolated nodes
varies dramatically (from 0 in DBLP to 15,877 in PubMed), providing systematic evaluation of how
methods handle nodes with no structural information. This diversity enables assessment of method
robustness across diferent bipartite graph characteristics while maintaining the common constraint
of feature-poor scenarios. Complete dataset statistics and experimental protocols are detailed in the
appendix.</p>
        <p>Our contribution addresses the gap between theoretical method development and practical
deployment guidance by providing systematic, tier-based evaluation that considers the full spectrum of
implementation and optimization constraints faced by practitioners working with feature-poor bipartite
graphs.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Empirical Evidence for Framework Hypotheses</title>
      <p>Our experimental evaluation across six bipartite graph datasets provides systematic evidence for the
ifve hypotheses underlying our method selection framework. The results demonstrate clear patterns
that validate our tier-based approach while revealing the practical trade-ofs that guide efective method
selection.</p>
      <sec id="sec-4-1">
        <title>5.1. Hypothesis 1: No Universal Winner</title>
      </sec>
      <sec id="sec-4-2">
        <title>5.2. Hypothesis 2: Implementation Maturity Efects</title>
        <p>The reliability challenges of Tier 4 methods provide strong evidence for implementation maturity efects.
VilLain and Subgradient methods failed to execute on multiple datasets despite theoretical sophistication,
with ”N/A” entries in Tables 3 and 4 indicating memory constraints that prevented execution. When
0.45
0.40</p>
        <p>CSP</p>
        <p>Tier1
Tier2
Tier3
Tier4</p>
        <p>HGCN_one_layer</p>
        <p>RF_features
HyperND</p>
        <p>HGCN
AllDeepSets</p>
        <p>MLP
AllSetTransformer</p>
        <p>GraphSAGE
AllSetTransformerNormalized</p>
        <p>AllDeepSetsNormalized
EHGNN</p>
        <p>VilLain</p>
        <p>HCHA
UniGCNII</p>
        <p>100 101
Average execution time over datasets [ms]
102
103
these methods did execute, they often underperformed simpler alternatives, suggesting that resource
requirements compromise their practical value.</p>
        <p>In contrast, well-established methods in Tiers 1-3 demonstrated consistent execution across all
datasets with reliable performance patterns. This reliability diference validates our framework’s
emphasis on implementation maturity alongside theoretical performance, supporting the progressive
tier structure that prioritizes proven methods before exploring experimental approaches.</p>
        <p>Even if Tier 4 method can achieve a competetive performance – VilLain in classification task – the
execution times one order of magnitude higher than other methods demonstrating downsite of their
not-mature implementation for practical use.</p>
      </sec>
      <sec id="sec-4-3">
        <title>5.3. Hypothesis 3: Dense Feature Extraction Trade-of</title>
        <p>Figures 1 and 2 dramatically illustrate the computational cost of dense feature extraction, showing 2–3
orders of magnitude diferences between Tier 1 methods (green) operating on one-hot representations
and higher tiers requiring dense features. The bottom-left clusters in both figures represent Tier 1
methods achieving competitive performance with minimal computational overhead.</p>
        <p>The performance-complexity trade-of varies by task and dataset. For retrieval tasks, several Tier 1</p>
        <p>VilLain
LR_features HGCN_one_layer</p>
        <p>HCHA</p>
        <p>MLP RF_features
LR_one_hot</p>
        <p>AllDeepSetsNormalized</p>
        <p>AllSetTransformerNormalized</p>
        <p>GraphSAGE</p>
        <p>AllDeepSets
UniGCNII</p>
        <p>AllSetTransformer
EHGNN</p>
        <p>100 101
Average execution time over datasets [ms]
102
103
methods achieve performance within 5% of the best while maintaining the fastest execution times.
Classification tasks show more pronounced benefits from dense representations, but often at
computational costs that may not justify modest performance improvements. This trade-of validates our tier
boundary between structural methods (Tier 1) and dense representation approaches (Tiers 2-3).</p>
      </sec>
      <sec id="sec-4-4">
        <title>5.4. Hypothesis 4: Simple Method Failure Principle</title>
        <p>When simple methods cannot reliably identify the most obvious positive examples in a dataset, complex
architectures rarely overcome these fundamental limitations. This principle provides a critical decision
point: rather than immediately progressing to sophisticated methods when Tier 1 performance appears
insuficient, practitioners should first assess whether the limitation stems from method inadequacy or
fundamental task characteristics that no learning approach can easily overcome.</p>
        <p>We operationalize this assessment through confident prediction performance, where  @100 proves
more suitable than accuracy for evaluating performance on ”simple examples”—instances where methods
should excel if they can capture fundamental data patterns. Our empirical analysis reveals that
highertier methods improve performance meaningfully only when Tier 1 methods achieve suficient confident
prediction performance. Based on our results,  @100 ≤ 0.5 serves as a practical indicator for this
threshold, though this value may require adjustment for diferent domains.</p>
        <p>This pattern is evident across our datasets: scenarios where even the strongest Tier 1 methods achieve
limited confident prediction performance (CiteSeer-CC at 0.503, Walmart at 0.286-0.350) consistently
show traditional methods matching or outperforming sophisticated GNNs. Conversely, datasets with
higher  @100 baselines (Cora variants, PubMed, DBLP with values above 0.7) demonstrate clearer
benefits from method progression, validating our Phase 1 assessment strategy.</p>
        <p>This hypothesis connects to our broader framework by providing a systematic decision point for
resource allocation. Rather than automatically progressing to higher tiers when Tier 1 performance
appears insuficient, practitioners should first assess whether the limitation stems from method inadequacy
or fundamental task characteristics that no learning approach can easily overcome.</p>
        <p>Confidence Assessment: We evaluate this hypothesis from multiple perspectives. The specific
threshold of 0.5 receives low-to-medium confidence, being based on only six datasets in our evaluation.
However, the  @ metric as an indicator of fundamental task failure appears quite reasonable for tasks
with reasonably uniform label distributions, where confident predictions should reflect the model’s
ability to identify the most obvious positive examples. The underlying assumption that test examples
range from easy to hard—while intuitively reasonable—lacks direct empirical confirmation in our study.
This assumption should be carefully considered when applying the framework, as violations could
explain cases where the framework’s guidance proves inefective. Future work should validate both
the threshold values and the easy/hard example assumption across broader domains to strengthen this
hypothesis.</p>
      </sec>
      <sec id="sec-4-5">
        <title>5.5. Hypothesis 5: Optimization Budget Constraints</title>
        <p>The consistent performance of Tier 1 and Tier 2 methods across both tasks provides evidence for
optimization budget efects. These methods benefit from manageable hyperparameter spaces that enable
thorough exploration within realistic time constraints. Table 2 shows Tier 2 methods achieving the
highest number of competitive performances (within 5% of best), suggesting that moderate complexity
with reliable optimization often outperforms sophisticated architectures with optimization uncertainty.</p>
        <p>Our time-boxed optimization approach reveals that Tier 3 methods with extensive configuration spaces
may not achieve their full potential within practical constraints. This limitation actually reinforces our
framework’s value: when simpler methods achieve competitive performance with reliable optimization,
the additional complexity and optimization uncertainty of sophisticated architectures becomes a practical
disadvantage rather than merely a theoretical concern.</p>
      </sec>
      <sec id="sec-4-6">
        <title>5.6. Integration of Evidence Supporting Framework Design</title>
        <p>The empirical evidence collectively validates our tier-based progression strategy. Tier 1 methods like
Label Propagation and CSP provide excellent starting points with minimal computational overhead
and reliable optimization. Tier 2 methods ofer systematic enhancement through dense representations
while maintaining manageable complexity. Tier 3 GNNs demonstrate sophisticated capabilities but with
optimization challenges that limit their practical advantages. Tier 4 methods show implementation
reliability issues that constrain their applicability.</p>
        <p>The dataset-dependent performance patterns, combined with clear computational and optimization
trade-ofs, support our systematic evaluation approach over universal method preferences. The
framework successfully identifies when simple methods sufice versus when sophistication is warranted,
providing practitioners with evidence-based decision points that balance performance requirements
against practical constraints.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Conclusions and Future Work</title>
      <p>Our systematic framework for method selection in feature-poor bipartite graphs provides
evidencebased guidance through comprehensive evaluation of 16 methods across six real-world datasets. We
established five empirical hypotheses about method efectiveness and validated a four-tier progression
strategy that balances performance requirements against computational constraints and implementation
complexity. Our results demonstrate that method efectiveness is fundamentally dataset-dependent, with
traditional approaches like Label Propagation often achieving competitive performance at 2–3 orders of
magnitude lower computational cost than Graph Neural Networks. The framework’s key insight is the
confident prediction assessment principle: when simple methods achieve poor performance on their
most confident predictions (roughly  @100 ≤ 0.5 ), complex architectures rarely provide significant
improvements, suggesting data quality enhancement as more efective than method sophistication. This
systematic approach enables practitioners to eficiently navigate the performance-complexity trade-of
space while avoiding over-engineering solutions for tasks where simple methods sufice.</p>
      <p>Limitations and Future Work: Our framework validation is based on six datasets primarily from
academic citation networks, requiring broader domain validation to strengthen generalizability claims.
The framework’s core assumption—that test examples contain a range of dificulties from very simple to
very hard—underlies our confident prediction assessment strategy and may not hold universally across
all domains. Future work should focus on systematic validation across diverse domains to identify
when this dificulty distribution assumption fails and how practitioners should adapt their approach
accordingly. The main challenge for framework adaptation lies in identifying appropriate metrics
for detecting ”simple examples” and establishing domain-specific thresholds for confident prediction
failure. For highly imbalanced tasks like threat detection, the  @100 ≤ 0.5 threshold would likely need
adjustment to much lower values, while other domains may require entirely diferent confidence metrics
to assess fundamental task tractability. Additionally, extending our tier-based evaluation protocol to
incorporate new architectures as they emerge will ensure the framework remains current and actionable
for practitioners facing feature-poor bipartite graph learning challenges.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The authors have not employed any Generative AI tools.</title>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>A. Implementation Details and Reproducibility</title>
      <p>This appendix provides essential implementation information for reproducing the presented results.
All methods were evaluated using identical experimental protocols with time-boxed hyperparameter
optimization to ensure fair comparison.</p>
      <sec id="sec-7-1">
        <title>A.1. Dataset Construction and Preprocessing</title>
        <p>Bipartite Graph Construction: Each dataset constructs bipartite graphs  = ( ,  , ) where 
represents primary entities (papers/products) and  represents secondary entities
(authors/citations/copurchases). Citation networks derive  from unique cited papers (CC variants) or individual authors
(CA variants). Walmart constructs  from co-purchase relationships.</p>
        <p>Feature Extraction: Matrix Factorization uses the implicit library with 50 iterations and confidence
weighting. Spectral Embedding performs eigenvalue decomposition using scipy.sparse.linalg.eigsh on
the normalized Laplacian. One-hot encoding uses incidence matrix rows directly.</p>
        <p>Data Splitting: Natural dataset splits when available, otherwise stratified random splitting ensuring
class balance. Training sets range from 53 nodes (PubMed) to 2,932 nodes (Walmart). Cross-validation
employed for limited training data scenarios.</p>
      </sec>
      <sec id="sec-7-2">
        <title>A.2. Method Implementation Sources</title>
        <p>Tier 1 Methods: CSP implemented via custom SQL operations. Label Propagation uses
sklearn.semi_supervised with bipartite-to-unipartite conversion. One-hot methods use sklearn.linear_model and
sklearn.naive_bayes directly on incidence matrix rows.</p>
        <p>
          Tier 2 Methods: Feature-based methods combine embedding techniques with sklearn
implementations. MLP uses implementation from [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] that adapted the implementation from [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. HGCN
single-layer implements a single HyperConv from the PyTorch Geometric library.
        </p>
        <p>
          Tier 3 Methods: HGCN multi-layer uses PyTorch Geometric implementation with 1-5 hidden layers.
GraphSAGE uses PyTorch Geometric implementation with previously applied bipartite-to-unipartite
conversion. HyperND, EHGNN, HCHA, UniGCNII, and AllSet variants use implementations provided
by [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ], the ED-HNN method proposed by was not included in results, because of consistently bad
performance with the current settings during the initial tests.
        </p>
        <p>
          Tier 4 Methods: VilLain uses an implementation from its authors [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. And subgradient method
uses a custom implementation without standard library support based on the provided by its authors
pseudocode algorithms [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], explaining execution failures on memory-constrained data
        </p>
      </sec>
      <sec id="sec-7-3">
        <title>A.3. Hyperparameter Configuration Spaces</title>
        <p>
          Traditional Methods: CSP is used in its parameter-free form [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Label Propagation searches alpha in
the range [0, 1.0] in 0.1 increments, num_layers [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1-5</xref>
          ]. Logistic Regression varies C [0.1, 1, 10], max_iter
[1000, 3000, 5000], solver [saga, liblinear, newton-cg]. Feature-based methods add embedding_dim [60,
120, 240, 480] and embedding type [mf, se], standing for matrix factorization and spectral embedding
correspondingly.
        </p>
        <p>
          Neural Methods: All neural methods utilize embeddings computed by non-negative matrix
factorization or spectral embedding methods with dimensionality in the range [32-512]. MLP adds layers
[
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1-5</xref>
          ], hidden size [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref14 ref15 ref16 ref17 ref18 ref19 ref20 ref21 ref22 ref23 ref24 ref8 ref9">8-1024</xref>
          ], dropout [0.0-0.8]. Multi-layer GNNs include layers [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1-5</xref>
          ], hidden [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref14 ref15 ref16 ref17 ref18 ref19 ref20 ref21 ref22 ref23 ref24 ref8 ref9">8-1024</xref>
          ],
dropout [0.0-0.8]. Advanced methods like EHGNN feature extensive configuration including
classiifer_hidden [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref14 ref15 ref16 ref17 ref18 ref19 ref20 ref21 ref22 ref23 ref24 ref8 ref9">8-256</xref>
          ], edconv_type [EquivSet, JumpLink, MeanDeg], activation [Id, relu, prelu]. VilLain
uses specialized parameters dim [128, 256, 512], steps [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5 ref6 ref7 ref8 ref9">1-9</xref>
          ], max_iter [
          <xref ref-type="bibr" rid="ref10">10, 100, 1000</xref>
          ].
        </p>
        <p>Optimization Protocol: Time-boxed random grid search with equal computational budget per
method. Neural methods use early stopping on validation performance. Traditional methods use
optimal hyperparameters directly. Configuration spaces range from tens of combinations (traditional)
to thousands (neural) – (see above), justifying the extensive optimization efort and explaining
computational cost diferences between tiers. We consider 1 hour time budget for each method/dataset for
ifnding best hyperparameter setup (evaluated on validation set).</p>
        <p>Computational Environment: All experiments conducted on identical hardware with consistent
memory and time constraints. The hardware specificaiton is Amazon EC2 G4.xlarge instances with
4 vCPUs, 16 GB RAM, and NVIDIA T4 GPU. All timing measurements exclude data loading and
preprocessing to focus on algorithm-specific computational costs. ”N/A” entries indicate memory
limitations preventing method execution rather than implementation failures.</p>
      </sec>
      <sec id="sec-7-4">
        <title>A.4. Results Presentation</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Procházka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dědič</surname>
          </string-name>
          , L. Bajer,
          <article-title>Convolutional signal propagation: A simple scalable algorithm for hypergraphs</article-title>
          ,
          <source>arXiv preprint arXiv:2409.17628</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Torr</surname>
          </string-name>
          ,
          <article-title>Hypergraph convolution and hypergraph attention</article-title>
          ,
          <source>Pattern Recognition</source>
          <volume>110</volume>
          (
          <year>2021</year>
          )
          <fpage>107637</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Tudisco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Benson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Prokopchik</surname>
          </string-name>
          ,
          <article-title>Nonlinear higher-order label spreading</article-title>
          ,
          <source>in: Proceedings of the Web Conference</source>
          <year>2021</year>
          ,
          <year>2021</year>
          , pp.
          <fpage>2402</fpage>
          -
          <lpage>2413</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Y.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shin</surname>
          </string-name>
          ,
          <article-title>Villain: Self-supervised learning on homogeneous hypergraphs without features via virtual label propagation</article-title>
          ,
          <source>in: Proceedings of the ACM Web Conference</source>
          <year>2024</year>
          ,
          <year>2024</year>
          , pp.
          <fpage>594</fpage>
          -
          <lpage>605</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. G.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Chan</surname>
          </string-name>
          ,
          <article-title>Re-revisiting learning on hypergraphs: confidence interval and subgradient method</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>4026</fpage>
          -
          <lpage>4034</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Schafer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Frankowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Herlocker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sen</surname>
          </string-name>
          ,
          <article-title>Collaborative filtering recommender systems, in: The adaptive web: methods and strategies of web personalization</article-title>
          , Springer,
          <year>2007</year>
          , pp.
          <fpage>291</fpage>
          -
          <lpage>324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>X.</given-names>
            <surname>ZhuЃ</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          <article-title>GhahramaniЃн, Learning from labeled and unlabeled data with label propagation, ProQuest number: information to all users (</article-title>
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Volinsky</surname>
          </string-name>
          ,
          <article-title>Matrix factorization techniques for recommender systems</article-title>
          ,
          <source>Computer</source>
          <volume>42</volume>
          (
          <year>2009</year>
          )
          <fpage>30</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Dai</surname>
          </string-name>
          , L. Song,
          <article-title>Discriminative embeddings of latent variable models for structured data</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>2702</fpage>
          -
          <lpage>2711</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Hamilton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <article-title>Inductive representation learning on large graphs</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Unignn: a unified framework for graph and hypergraph neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:2105.00956</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <article-title>Edge representation learning with hypergraphs</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>34</volume>
          (
          <year>2021</year>
          )
          <fpage>7534</fpage>
          -
          <lpage>7546</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Torr</surname>
          </string-name>
          ,
          <article-title>Hypergraph convolution and hypergraph attention</article-title>
          ,
          <source>Pattern Recognition</source>
          <volume>110</volume>
          (
          <year>2021</year>
          )
          <fpage>107637</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>E.</given-names>
            <surname>Chien</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Milenkovic</surname>
          </string-name>
          ,
          <article-title>You are allset: A multiset function framework for hypergraph neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:2106.13264</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chu</surname>
          </string-name>
          ,
          <article-title>Automl: A survey of the state-of-the-art, Knowledge-Based Systems 212 (</article-title>
          <year>2021</year>
          )
          <fpage>106622</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Y.</given-names>
            <surname>Philip</surname>
          </string-name>
          ,
          <article-title>A comprehensive survey on graph neural networks</article-title>
          ,
          <source>IEEE transactions on neural networks and learning systems 32</source>
          (
          <year>2020</year>
          )
          <fpage>4</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Sculley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Holt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Golovin</surname>
          </string-name>
          , E. Davydov,
          <string-name>
            <given-names>T.</given-names>
            <surname>Phillips</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ebner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Crespo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dennison</surname>
          </string-name>
          ,
          <article-title>Hidden technical debt in machine learning systems</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>28</volume>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>You</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <article-title>Design space for graph neural networks</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>17009</fpage>
          -
          <lpage>17021</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>A. K. McCallum</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Nigam</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rennie</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Seymore</surname>
          </string-name>
          ,
          <article-title>Automating the construction of internet portals with machine learning</article-title>
          ,
          <source>Information Retrieval 3</source>
          (
          <year>2000</year>
          )
          <fpage>127</fpage>
          -
          <lpage>163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>C. L.</given-names>
            <surname>Giles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Bollacker</surname>
          </string-name>
          , S. Lawrence,
          <string-name>
            <surname>Citeseer:</surname>
          </string-name>
          <article-title>An automatic citation indexing system</article-title>
          ,
          <source>in: Proceedings of the third ACM conference on Digital libraries</source>
          ,
          <year>1998</year>
          , pp.
          <fpage>89</fpage>
          -
          <lpage>98</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>P.</given-names>
            <surname>Sen</surname>
          </string-name>
          , G. Namata,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bilgic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Getoor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Galligher</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          Eliassi-Rad,
          <article-title>Collective classification in network data</article-title>
          ,
          <source>AI</source>
          magazine
          <volume>29</volume>
          (
          <year>2008</year>
          )
          <fpage>93</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Yao,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <article-title>Arnetminer: extraction and mining of academic social networks</article-title>
          ,
          <source>in: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '08,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2008</year>
          , p.
          <fpage>990</fpage>
          -
          <lpage>998</lpage>
          . URL: https://doi.org/10.1145/1401890.1402008. doi:
          <volume>10</volume>
          .1145/1401890.1402008.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Equivariant hypergraph difusion neural operators</article-title>
          ,
          <source>arXiv preprint arXiv:2207.06680</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-N.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Benson</surname>
          </string-name>
          ,
          <article-title>Combining label propagation and simple models out-performs graph neural networks</article-title>
          , arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>13993</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>