<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Responsible Data Management for Human Resources</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dimitrios Vogiatzis∗∗</string-name>
          <email>dimitrv@acg.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olivia Kyriakidou</string-name>
          <email>OKyriakidou@acg.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>The American College of Greece</institution>
          ,
          <addr-line>Deree</addr-line>
          ,
          <institution>&amp; NCSR "Demokritos"</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The American College of Greece</institution>
          ,
          <addr-line>Deree, Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <abstract>
        <p>The human resources (HR) departments rely increasingly on recommender systems (RS) for most of their processes, such as recruiting, selecting and developing their employees. However, the RS often discriminate unfairly based on biases in data that may perpetuate and enhance existing biases and in the work place. An important part of an HR department is a the data ecosystem, comprising raw and derived data, related to potentially diferent stakeholders while being subject to laws, and regulations. In this work we propose the characteristics of a data ecosystem that will facilitate data transparency through traceability as a way of detecting potential biases in the data. • Information systems → Data management systems; • Social and professional topics → Employment issues; User characteristics.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
    </sec>
    <sec id="sec-2">
      <title>INTRODUCTION</title>
      <p>Recommender systems (RS) are widely used by Human Resources
(HR) departments to facilitate their business processes from the
point of time optimization, but also from the perspective of
minizing human intervention, in an efort of achieving fairness. RS can
be applied in the recruitment, hiring, promotion of employees, etc.
For instance they can used in matching CVs against job posts, in
ranking CVs which determines the order of interviews or even in
comparison of CVs against past CVs of employees that are deemed
as successful. A RS can also be used for segmenting job
applications into categories, detecting and recording long terms trends etc.
Moreover, they can be used by prospective employees that seek
employment.</p>
      <p>Although RS seem to remove human intervention by automating
HR processes, often but not exclusively, through advanced machine
learning algorithms, still segments of the population can be
discriminated against. The data upon which the analysis is based on
may contain biases towards age groups, gender, ethnic origin etc.
The data stem from specific data items, specific data features, data
distributions, data sampling methods. The bias in data can be also
very subtle and dificult to detect as it may appear in derived data
stemming from an analytics process.
∗Both authors contributed equally to this research.</p>
      <p>Eventually a RS in HR is a information system that is directly
related to the professional life of people, and as such it should be
subject to ethical and legal regulations, apart from the technical
ones, like prediction accuracy. ACM 1 and IEEE 2 have issued codes
of ethics that refer to the need for fairness in Information Systems.
In particular section 1.4 of ACM code of ethics is entitled Be fair
and take action not to discriminate and section II of the IEEE code
of ethics states: To treat all persons fairly and with respect, to not
engage in harassment or discrimination, and to avoid injuring others.</p>
      <p>
        In reality RS are often fraught with elements of discrimination
and unfairness. The unfairness may stem from biases in the data
that may misrepresent the actual population and subsequent
analytic algorithm often amplify the data biases. Lack of fairness can
have potential legal consequences, especially in employment as it
might violate anti-discrimination laws. Also it might have financial
consequences as usage of such systems might drop. See also [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
for a recent tutorial on the origin and form of fairness in RS.
      </p>
      <p>
        A data ecosystem is a network of data in potentially many forms
(e.g. unstructured, structured) as well as accompanying rules that
permit their acquisition, storage, maintenance, and retrieval. The
data ecosystems is of potential interest to many stakeholders,
including data providers, and data users that try to create value out
the ecosystems. A data ecosystem includes metadata, as well as
legal, organizational or ethical regulations. Moreover, the ecosystems
evolve as their constituent components change. Finally, derived
data also form part of the ecosystems. For instance clusters,
predictions etc. are examples of derived data that are produced by
statistical or machine learning methods. See [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] for an overview
of data ecosystems.
      </p>
      <p>
        Our contribution is to focus on the data component of a RS and
examine how a data ecosystem would facilitate data transparency
through data traceability so that potential biases are made explicit
or a least easier to track and detect. Our approach is based on a
similar work for data transparency in the biomedical domain [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
2
      </p>
    </sec>
    <sec id="sec-3">
      <title>RELATED WORK</title>
      <p>
        Responsible data management has been discussed in the context
of automated decision systems (ADS), which are systems that
make make decisions about humans that might afect their
socioeconomic life [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. They authors refer to the ethical challenges
faced in all phases of a data science pipeline and the need for a fair,
transparent and responsible data management.
      </p>
      <p>Our current work focuses on the data representation part of
an ADS, which is essentially an RS. In particular, we refer to the
features of a data ecosystem and how it can be supported with
1https://www.acm.org/code-of-ethics
2https://www.ieee.org/about/corporate/governance/p7-8.html
semantic web technologies, and their relevance to the issues of
fairness in HR.</p>
      <p>A very similar approach that we propose in the current work
has been developed for a biomedical system in the context of the
EU funded project BigMedilytics 3 for a lung-cancer pilot
application. The pilot integrates structured and unstructured information,
open and sensitive data in a knowledge graph. This constitutes an
example of a data ecosystem.
3</p>
    </sec>
    <sec id="sec-4">
      <title>MOTIVATING EXAMPLES</title>
      <p>Next we mention some examples that indicate the form of bias in
raw data, in data associations, as well as in derived data produced by
machine learning algorithms. The examples refer to HR department
cases.</p>
      <p>
        Biased Data based on human behavioral biases. HR algorithmic
recommendations may sustain existing inequities when they are
trained on data that do not include specific groups of individuals [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
For example, many selection algorithms try to identify the criteria
that characterize the ideal employee and use them for the selection
of newcomers. For this task they utilize performance data that
identify the best performing employees within the organize and
then identify the traits that distinguish them. However, there is the
danger that if the performance data favor men due to existing biases
within the organization [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], then the selection algorithm might
include gender as a preferred characteristic for the ideal candidate
and prefer men rather than women applicants. In this sense, existing
biases could be reified by limiting the number of certain groups and
possibly underrepresented groups who are alerted, selected, and
hired for specific job openings [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Moreover, HR recommendation
systems utilized for the automated screening of candidates’ CVs
against certain preferred selection criteria may also generate biased
results when they are trained on data from past hiring decisions that
are based on individual, organizational and structural biases against
certain underrepresented groups of employees [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The use of
natural language processing (NLP) tools in chatbots that evaluate
candidates’ competencies and fit to the job and the organization
may also preserve existing societal inequities when they are trained
on biased data and exclude certain categories of candidates. The
association of African-American names with negative feelings and
female names with the household and non-technical jobs has been
already documented in the literature [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>
        Proxies. Recommendation systems can replicate biases in other
subtle ways, especially through the use of proxies. Certain hiring
criteria could serve as proxies for categorizing individuals in specific
groups and drive discrimination. For example, the use of gaps in
employment as a hiring criterion could discriminate against women
applicants as women disproportionately leave the workplace to
provide child or elderly care [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Moreover, job matching platforms
and job recommendation systems use proxies for “relevance” that
reproduce biases. Such systems, for example, could show to women
specific jobs at specific hierarchical levels (e.g., senior or junior
positions in management) according to their own search history
but also according to the search history of women similar to them.
Accordingly, they might end up with fewer recommendations for
3https://www.bigmedilytics.eu/
senior positions if themselves and others look tend to look for
lower-level jobs [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Proxies are also included in HR data that train
employee selection recommendation systems in order to ofer the
most appropriate renumeration package to prospective employees.
Such suggestions however may reinforce gender or racial pay gaps
especially when they reflect the existence of strong proxies that
signal for certain gender representations (e.g., male employees as
breadwinners) and status inequalities.
      </p>
      <p>
        Facial analysis that is used in virtual interviews may also create
disparate impact on specific sub-groups of employees across gender
and racial lines. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] it was shown that the faces of women with
darker skin cannot be reliably recognized by facial analysis systems
as well as the emotions of people with disabilities and in diferent
cultural contexts [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Finally, in employee selection, most recruiters
use a number of candidate characteristics as proxies of culture fit [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
defined as the degree to which the values of the individual match
those of the organization. However, there is the danger that these
proxies will become hard rules ignoring their subjective character
and in this way exclude certain individuals who are thought apriori
that they do not “fit” the organizational culture.
      </p>
      <p>
        Segregation of individuals. Biases could also persist when
algorithms segregate employees into groups drawing inferences about
individuals from their group memberships. Selection
recommendation systems, for instance, may erroneously attribute to people
with disabilities [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] certain characteristics based on their group
membership without properly assessing the candidates and
consequently ofer lower status job positions. Moreover, categorizing
individuals into certain gender groups could unfairly marginalize
non-binary and transgender employees while their classification
into certain race groups could signify status inequalities [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
Human computer interaction. Most HR recommendation systems
run on platforms that require employees’ and candidates’ active
involvement with them which is determined merely by the rules
set by the platform that control all processes [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. For instance, job
candidates do not have any control over how their application will
be presented to possible employers and they have to provide all
the required information by the platform if they want to be
considered for future job opportunities. Moreover, employee selection
recommendation systems tend to present numerical rankings of
candidates to employers generating the perception that there are
actual substantial diferences between the candidates for a certain
position, while in reality the diferences might be minimal [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
4
      </p>
    </sec>
    <sec id="sec-5">
      <title>REQUIREMENTS FOR A DATA ECOSYSTEM IN HR</title>
      <p>Next, based on the previously mentioned examples we sketch the
requirements that would be necessary for a data ecosystem in HR.</p>
      <p>Data management requirements: The data ecosystem should
allow data sharing for structured (e.g. CSV files) and
unstructured data (e.g. text). The data should be accessible and
retrievable by all stakeholders. Also the data has to be of high
quality. For instance, data items that have missing values
could be rejected or data items that are very old. As an
example we could mention a CV that that does not contain any
information about education or past employment. The data
management requirements fall into the following categories:</p>
      <sec id="sec-5-1">
        <title>DM1: Data management of multiple document types</title>
        <p>should be supported. DM2: Quality of data items should
be supported at all levels of data pipeline, e.g. at the raw data,
but also at derived data.</p>
        <p>Organizational requirements: The data should be stored,
accessed and processed according to the organization’s rules,
and regulations. The organization requirements fall into the
following categories: O1: Data governance should be
enforced by the organization. Thus the data acquisition process,
the data storage, and retention, access rights, and data
obsolescence are items related to data governance. The HR
department may have business rules that stipulate the
recruitment policy, and what will be the requested documents.
O2: Data sovereignty which specifies who owns the
original data, the derived data, and to what purpose. This will
increase the trust in the system. For instance, it will be clearer
of how a submitted CV will be handled.</p>
      </sec>
      <sec id="sec-5-2">
        <title>Legal &amp; Ethical requirements: The data management should</title>
        <p>be in accordance with the requirements of the European
GDPR. 4 Moreover, the data management should address
bias. For instance, the execution of algorithms should be
independent of sensitive attributes (like ethnicity, age, gender).
In addition, the data should be owned and used for the
indented purposes. For instance, CVs of job applicants should
not used to generate business value by selling them without
their consent. Finally, traceability is an import aspect of the
data that essentially allows to know where the data where
obtained from, and how they were obtained.</p>
        <p>The above can be summarized into the following ethical
requirements: E1: Data protection &amp; ownership which
specifies the extence of ownership for each stake holder.
E2: Sensitive attributes which clearly states the sensitive
attributes, with the foresight that they should be used by
prediction algorithms. Typically, they represent age, gender,
ethnic background etc. The sensitive attributes are typically
associated with the provisions of GDPR.</p>
      </sec>
      <sec id="sec-5-3">
        <title>E3: discrimination attributes they may lead to discrimi</title>
        <p>nation in non-obvious ways. For instance the name of a job
applicant might inadvertently facilitate discrimination as it
may reveal ethnic origin. Moreover, some derived attributes
fall in this category, for instance employment gaps.
5</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>DESIGN OF A DATA ECOSYSTEM</title>
      <p>
        We present in detail the concept of a data ecosystem, that will
serve as the infrastructure for an HR. A data ecosystem (DE) can
be defined as a 4-tuple: DE=&lt;Data Sets, Data Operators, Meta-Data,
Mappings&gt; [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Data sets: the ecosystem is composed of potentially multiple
data sets. Data sets can be comprised of structured or
unstructured information; also, they have diferent formats, e.g.,
CSV, JSON or tabular relations, and can be managed using
diferent management systems.
4https://gdpr-info.eu/
Data operators: the set of operators that can be executed
against the data sets. For instance, anonymization, data
quality checks, recency checks can be considered as data
operators.</p>
      <p>Meta-Data: provide the semantics of the data stored in the
data sets of the data ecosystem. It comprises:
(1) A Domain ontology, which provides a unified view of the
concepts, relationships, and constraints of the domain of
knowledge. It associates formal elements from the domain
ontology to concepts. For instance, a specific job post and
a specific applicant can be part of the concepts in a domain
ontology.
(2) Properties enable the definition of data quality, provenance,
and data access regulations of the data in the ecosystem.
For instance, last updated and other non-domain
properties (quality etc).
(3) Descriptions of the main characteristics of a data set. No
specific formal language or vocabulary is required; in fact,
a data set could be described using natural language. For
instance, Data set D is a collection of CVs and cover letters.
Mappings expressing correspondences among the diferent
components of a data ecosystem. The mappings are as
follows:</p>
      <sec id="sec-6-1">
        <title>Mappings between ontologies: they represent associations</title>
        <p>between the concepts in the diferent ontologies that
compose the domain ontology of the ecosystem. For instance, if
there can a mapping between the personnel ontology, and
the candidate employees ontology.</p>
        <p>Mappings between data sets: they represent relations among
data sets of the ecosystem and the domain ontology.
6</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>DATA ECOSYSTEM IN HR</title>
      <p>The role of a data ecosystem (DE) is to provide an explicit
description of the data and applicable operations on them through
metadata and mapping rules. Next we provide some examples that refer
to the usage of the elements of the DE as referred to in the previous
section. The description of a data ecosystem that we describe next
does not cover all the activities HR, but rather it addresses some
essential parts that refer to recruitment and hiring of employees.
Thus we will assume a scenario where there are applicants CVs and
job posts. The DE for this example can be depicted in Figure 1.</p>
      <p>One of the data sources are the CVs of the applicants. Typically,
they contain textual information, possibly with some keywords (e.g.
education, past employment) which can be helpful as annotations.
Thus a CV represents a piece of unstructured or partially structured
information.</p>
      <p>A Data Operator can implement an NLP process to extracte
structured information from the CV. Typically named entity recognition
(NER) and relation extraction (RE) will have to be performed,
resulting in triplets comprising two entities and a relation. The named
entities (NE) in CVs can be things like: skills, past employment,
educational achievements and demographic data. The relations
connect the NE to the person in question, while being labeled with
time annotations. This will form the job applicant’s graph. The
extracted entities will then be annotated with meta data derived
from a domain ontology, commonly described in OWL. For example
the NAICS 5 can be used to characterise the entities that refer to
the industry of employment.</p>
      <p>The NER and RE processes can on occasion be of low
precision. The Meta-data properties can represent the quality of the NLP
process as a numerical score per NE or per relation.</p>
      <p>To the best of our knowledge there is not single ontology that
is complete enough to annotate a CV for the requirements of an
HR. For instance, it may be necessary apart from NAICS to use
also resumeRDF, 6 and the Human Resources Ontology. 7 This
results in the need also to have mappings between the ontologies
for the common concepts (i.e. classes) and for the common object
properties. The mapping rules can be stated in RML. 8</p>
      <p>The second major source of information is the job post, which
is typically in textual form, possibly split in sections each with a
meaningful keyword (like company culture, required skills etc.).
This usually constitutes a partially structured piece of information.
Likewise with the case of CVs information has to be extracted in
the forms of triplets, resulting in the job posts graph. However,
it may be not necessary to extract structure from all parts of the
document. For instance a company’s culture could fall under the
Meta-Data Descriptions data set.</p>
      <p>Finally the merging of the two graphs in the integrated
knowledge graph can also be achieved with mapping rules. The mapping
rules, as well all the ontology selection, and possibly expansion
to be designed in cooperation of a knowledge engineer with a
representative of HR department.</p>
      <p>The issue of detecting possible bias in the data can be assisted
through data transparency, especially at the stage of NE annotation.
Thus CV attributes can be split into sensitive and non-sensitive
ones. The former comprising name, gender, ethnic origin, age etc.
whereas the latter would comprise entities like education, or skills.
5North American Industry Classification System
https://www.census.gov/naics/
6http://rdfs.org/resume-rdf/
7https://github.com/motapinto/cv-ontology/blob/main/cv-onto logy.owl
8https://rml.io/specs/rml/
(NAICS)</p>
      <p>The distinction between attributes can be represented for instance
as classes by expanding one of the existing ontologies. Thus is will
be clearer what attributes should be used from subsequent machine
learning algorithms that perform job recommendations.</p>
      <p>Subsequently a similar distinction can be made between soft and
hard skills in job posts. Naturally, this will require Data operators
to split the skills into two classes. This will facilitate an association
of soft and hard skills to the level of seniority of the position, and
to the applicants’ gender. This can reveal subtle forms of biases.</p>
      <p>Finally, business regulations and regulations derived from
ethical data management can be set as constraints on the integrate
knowledge graph, and be expressed in the SHACL 9 language.</p>
      <p>Typically the DE can accessed via SPARQL endpoints. Normally,
the end user has access via web services accessible through a
dashboard. The web services can also allow for diferent user roles, thus
implementing data access control.
7</p>
    </sec>
    <sec id="sec-8">
      <title>CONCLUSIONS</title>
      <p>In the current work we proposed a framework for responsible data
management for a human resources department. The framework is
based on the concept of a DE, that comprises data, meta-data and
data operators. It can be implemented with semantic technologies
(RDF Schema, OWL, RML rules, etc.). The implementation of a
data ecosystem will require a substantial investment both from a
knowledge engineering and the HR perspective. The benefits can
be important, especially in the field of data transparency. Moreover,
a DE can also facilitate the deployment of explainable machine
learning algorithms.</p>
    </sec>
    <sec id="sec-9">
      <title>ACKNOWLEDGMENTS</title>
      <p>The authors would like to acknowledge the support of the Deree
The American College of Greece in the current article.
9https://www.w3.org/TR/shacl/</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Ifeoma</given-names>
            <surname>Ajunwa</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>The Paradox of Automation as Anti-Bias Intervention</article-title>
          ,
          <volume>41</volume>
          Cardozo, L.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Ifeoma</given-names>
            <surname>Ajunwa</surname>
          </string-name>
          and
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Greene</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Platforms at work: Automated hiring platforms and other new intermediaries in the organization of work. In Work and labor in the digital age</article-title>
          .
          <source>Emerald Publishing Limited.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Lisa</given-names>
            <surname>Feldman</surname>
          </string-name>
          <string-name>
            <given-names>Barrett</given-names>
            , Ralph Adolphs, Stacy Marsella,
            <surname>Aleix M Martinez</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Seth D</given-names>
            <surname>Pollak</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Emotional expressions reconsidered: Challenges to inferring emotion from human facial movements</article-title>
          .
          <source>Psychological science in the public interest 20</source>
          ,
          <issue>1</issue>
          (
          <year>2019</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>68</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Miranda</given-names>
            <surname>Bogen</surname>
          </string-name>
          and
          <string-name>
            <given-names>Aaron</given-names>
            <surname>Rieke</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Help wanted: An examination of hiring algorithms, equity, and bias</article-title>
          . (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Joy</given-names>
            <surname>Buolamwini</surname>
          </string-name>
          and
          <string-name>
            <given-names>Timnit</given-names>
            <surname>Gebru</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Gender shades: Intersectional accuracy disparities in commercial gender classification</article-title>
          . In Conference on fairness,
          <source>accountability and transparency. PMLR</source>
          ,
          <fpage>77</fpage>
          -
          <lpage>91</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Robin</given-names>
            <surname>Burke</surname>
          </string-name>
          , Nasim Sonboli, and
          <string-name>
            <surname>Aldo</surname>
          </string-name>
          Ordonez-Gauger.
          <year>2018</year>
          .
          <article-title>Balanced neighborhoods for multi-sided fairness in recommendation</article-title>
          . In Conference on Fairness,
          <source>Accountability and Transparency. PMLR</source>
          ,
          <fpage>202</fpage>
          -
          <lpage>214</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Hege</surname>
            <given-names>H Bye</given-names>
          </string-name>
          , Henrik Herrebrøden, Gunnhild J Hjetland,
          <source>Guro Ø Røyset, and Linda L Westby</source>
          .
          <year>2014</year>
          .
          <article-title>Stereotypes of Norwegian social groups</article-title>
          .
          <source>Scandinavian Journal of Psychology</source>
          <volume>55</volume>
          ,
          <issue>5</issue>
          (
          <year>2014</year>
          ),
          <fpage>469</fpage>
          -
          <lpage>476</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Cinzia</given-names>
            <surname>Capiello</surname>
          </string-name>
          , Avigdor Gal, Matthias Jarke, and
          <string-name>
            <given-names>Jakob</given-names>
            <surname>Rehof</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Data ecosystems: sovereign data exchange among organizations (Dagstuhl Seminar 19391)</article-title>
          .
          <source>In Dagstuhl Reports</source>
          , Vol.
          <volume>9</volume>
          .
          <string-name>
            <surname>Schloss</surname>
          </string-name>
          Dagstuhl-Leibniz-Zentrum fuer Informatik.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Kate</given-names>
            <surname>Crawford</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>The hidden biases in big data</article-title>
          .
          <source>Harvard business review 1</source>
          ,
          <issue>4</issue>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Michael</surname>
            <given-names>D</given-names>
          </string-name>
          <string-name>
            <surname>Ekstrand</surname>
            ,
            <given-names>Robin</given-names>
          </string-name>
          <string-name>
            <surname>Burke</surname>
            , and
            <given-names>Fernando</given-names>
          </string-name>
          <string-name>
            <surname>Diaz</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Fairness and discrimination in recommendation and retrieval</article-title>
          .
          <source>In Proceedings of the 13th ACM Conference on Recommender Systems</source>
          .
          <volume>576</volume>
          -
          <fpage>577</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Sandra</surname>
            <given-names>Geisler</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maria-Esther</surname>
            <given-names>Vidal</given-names>
          </string-name>
          , Cinzia Capiello, Bernadette Farias Loscio, Avigdor Gal, Matthias Jarke, Maurizio Lenzerini, Paolo Missier, Boris Otto,
          <string-name>
            <given-names>Elda</given-names>
            <surname>Paja</surname>
          </string-name>
          , et al.
          <year>2021</year>
          .
          <article-title>Knowledge-driven Data Ecosystems Towards Data Transparency</article-title>
          .
          <source>arXiv preprint arXiv:2105.09312</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Os</given-names>
            <surname>Keyes</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>The misgendering machines: Trans/HCI implications of automatic gender recognition</article-title>
          .
          <source>Proceedings of the ACM on human-computer interaction 2</source>
          ,
          <string-name>
            <surname>CSCW</surname>
          </string-name>
          (
          <year>2018</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Karen</given-names>
            <surname>Levy</surname>
          </string-name>
          and
          <string-name>
            <given-names>Solon</given-names>
            <surname>Barocas</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Designing against discrimination in online markets</article-title>
          .
          <source>Berkeley Technology Law Journal</source>
          <volume>32</volume>
          ,
          <issue>3</issue>
          (
          <year>2017</year>
          ),
          <fpage>1183</fpage>
          -
          <lpage>1238</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Kirsten</given-names>
            <surname>Martin</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Ethical implications and accountability of algorithms</article-title>
          .
          <source>Journal of Business Ethics</source>
          <volume>160</volume>
          ,
          <issue>4</issue>
          (
          <year>2019</year>
          ),
          <fpage>835</fpage>
          -
          <lpage>850</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Marcelo</given-names>
            <surname>Iury S Oliveira and Bernadette Farias Lóscio</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>What is a data ecosystem?</article-title>
          .
          <source>In Proceedings of the 19th Annual International Conference on Digital Government Research: Governance in the Data Age. 1-9.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Lauren</surname>
            <given-names>A</given-names>
          </string-name>
          <string-name>
            <surname>Rivera</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Go with your gut: Emotion and evaluation in job interviews</article-title>
          .
          <source>American journal of sociology 120</source>
          ,
          <issue>5</issue>
          (
          <year>2015</year>
          ),
          <fpage>1339</fpage>
          -
          <lpage>1389</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Julia</surname>
            <given-names>Stoyanovich</given-names>
          </string-name>
          , Bill Howe, Serge Abiteboul, Gerome Miklau, Arnaud Sahuguet, and
          <string-name>
            <given-names>Gerhard</given-names>
            <surname>Weikum</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Fides: Towards a platform for responsible data science</article-title>
          .
          <source>In Proceedings of the 29th International Conference on Scientific and Statistical Database Management. 1-6.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Julia</surname>
            <given-names>Stoyanovich</given-names>
          </string-name>
          ,
          <source>Bill Howe, and HV Jagadish</source>
          .
          <year>2020</year>
          .
          <article-title>Responsible data management</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>13</volume>
          ,
          <issue>12</issue>
          (
          <year>2020</year>
          ),
          <fpage>3474</fpage>
          -
          <lpage>3488</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Adam</given-names>
            <surname>Sutton</surname>
          </string-name>
          , Thomas Lansdall-Welfare, and
          <string-name>
            <given-names>Nello</given-names>
            <surname>Cristianini</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Biased embeddings from wild data: Measuring, understanding and removing</article-title>
          .
          <source>In International Symposium on Intelligent Data Analysis</source>
          . Springer,
          <fpage>328</fpage>
          -
          <lpage>339</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Shari</given-names>
            <surname>Trewin</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>AI fairness for people with disabilities: Point of view</article-title>
          . arXiv preprint arXiv:
          <year>1811</year>
          .
          <volume>10670</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>