<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Lightning Talk: Making It Easier To Understand Research Software Impact</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Neil P. Chue Hong Software Sustainability Institute University of Edinburgh JCMB</institution>
          ,
          <addr-line>Peter Guthrie Tait Road Edinburgh, EH9 3FD</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          <addr-line>ORCID: 0000-0002-8876-7606</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>- How can we make it easier to understand research software impact? Is it easy for new researchers to start research in this area? In particular, are the tools available that would let them generate the research that we as the research software community require to convince our funders, fellow researchers and the public at large that having sustainable, open-use research software is important? 1 Depsy: http://depsy.org/ 2 Software Discovery Dashboard: https://github.com/mozillascience/software-discovery-dashboard 3 CodeMeta: https://github.com/codemeta/codemeta 4 Code as a Research Object: https://science.mozilla.org/projects/codemeta 5 ContentMine: http://contentmine.org/ 6 ScholarNinja: https://github.com/ScholarNinja/software_metadata 7 Lagatto: https://github.com/lagotto/lagotto 8 Libraries.io: https://libraries.io/ 9 SciCrunch: https://scicrunch.org 10 Resource Identification Initiative: https://www.force11.org/group/resource-identification-initiative 11 CodeMeta Participants: http://codemeta.github.io/</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Index Terms — software, research software, software metrics,
software impact.</p>
    </sec>
    <sec id="sec-2">
      <title>I. BACKGROUND</title>
      <p>In this lightning talk I will pitch the idea of the WSSSPE
community working together to define a conceptual framework
and architecture for the tools that are useful to measure
research software impact for instance:
- mining scholarly literature for mentions of software
- identifying the linkages between pieces of software
- visualising dependencies between pieces of software
or between software and other artefacts</p>
      <p>There are already a number of tools that have been
developed by others. Depsy1 is a tool which mines papers to find
fulltext mentions of software (currently only those written in
Python and R) and analyses GitHub repositories to see where
software is being used, thus giving a measure of the impact of a
piece of research software or library. The Software Discovery
Dashboard2 aims to search multiple code hosting services, such
as Zenodo, Figshare, and GitHub, for scientific software and
undertake analysis of it, utilising the Codemeta3 metadata
standards for describing scientific software and building on
previous work by Mozilla Science Lab, Github and Figshare on
Code as a Research Object4. ContentMine5 provides a platform
This work is licensed under a CC-BY-4.0 license.
for machine mining academic papers. ScholarNinja6 provided a
way of scraping software metadata. Lagatto7 is a tool for
tracking analysing article level events in scholarly content which has
been used to identify GitHub repositories in scientific papers.
Libraries.io8 provides a way identifying open source library
dependencies and discovering alternative software. SciCrunch9
is developing ways of allocating resource identifiers to
software using Research Resource Identifiers (RRIDs)10 to improve
citation and discovery.</p>
      <p>What I propose is a working group in WSSSPE to:
- Collectively capture all the useful tools that the
community is aware of
- Summarise the use cases for tools in this area
- Define a conceptual framework / architecture
- Identify any areas of functionality which are missing
- Determine what we should work on first
- Run hackathons / sprints to address low-hanging fruit
As researchers, we use evidence and experiences to test our
hypotheses and inform our work. However we are lacking the
tools to make this easy in the area of research software. By
working together, we can ensure that our efforts are aligned.</p>
    </sec>
    <sec id="sec-3">
      <title>ACKNOWLEDGMENT</title>
      <p>The Software Sustainability Institute is supported by the
EPSRC, ESRC and BBSRC through grant EP/N006410/1.
This paper is based on input from many members of the
Software Sustainability Institute team, Fellows, and
collaborators including the participants of the CodeMeta
workshop11, especially Abigail Cabunoc-Reyes, James
Howison and Daniel S. Katz.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>