Scrapers support discovering software repositories from different resources that might then be added via a core parser. By default, a scraper will query for some number of latest updated entries. If you provide a query string, it will be searched for that instead.

How does it work?

It’s very common to find software cross-referenced, meaning that you might find a GitHub repository represented in a Journal of Open Software paper, or a random database of software having mentions of both. In fact, we can look at these resources as databases that are regularly updated with research software that we want to add from. Since the research software encyclopedia is optimized to be maintained as a repository flat-file “database” on version control (GitHub), we have chosen to honor version control software as our bases of truth. For example, although we might find metadata in JoSS, since it’s typically pointing back to code in a version control system, we just use JoSS to find these repositories. If we find additional valuable metadata, however, (e.g., a DOI) we add this to the base GitHub or other version control record. We use this strategy as a best effort to not have redundancy in the representation of software repositories, but still capture metadata that we need.

Scrapers

The following scrapers are planned for development, depending on if an API is available, and it includes links to repositories:

Journal of Open Source Software

The Journal of Open Source Software is one of my favorite resources for new software. Although the site has an RSS (atom) feed, you can’t easily retrieve GitHub urls from it, so instead we parse the main site, either for latest, or based on a search query:

$ rse scrape joss
INFO:rse.main.scrapers.joss:Found repository: https://github.com/LiberTEM/LiberTEM
INFO:rse.main.scrapers.joss:Found repository: https://github.com/samhforbes/PupillometryR
INFO:rse.main.scrapers.joss:Found repository: https://github.com/Alcampopiano/hypothesize
INFO:rse.main.scrapers.joss:Found repository: https://github.com/BartoszBartmanski/StoSpa2
INFO:rse.main.scrapers.joss:Found repository: https://github.com/rgmyr/corebreakout
INFO:rse.main.scrapers.joss:Found repository: https://github.com/coljac/sensie
INFO:rse.main.scrapers.joss:Found repository: https://github.com/HajkD/LTRpred
INFO:rse.main.scrapers.joss:Found repository: https://github.com/mzy2240/ESA
INFO:rse.main.scrapers.joss:Found repository: https://github.com/KVSlab/turtleFSI.git
INFO:rse.main.scrapers.joss:Found repository: https://github.com/anmolter/XLUR

If you don’t want to add repositories to the database (but just do a test run) add --dry-run

$ rse scrape --dryrun joss

You can also scrape based on a term of interest:

$ rse scrape --dryrun joss docker

Within Python

And of course you can interact with a scraper from within Python! Either of the following will work to get the joss scraper:

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('joss')
from rse.main.scrapers import JossScraper
scraper = JossScraper()

And then you can either search for a term, or get the latest (on the front page)

results = scraper.latest()
results = scraper.search("docker")

By default, the latest won’t paginate (it would parse all of JoSS) but you can force it to:

results = scraper.latest(paginate=True)

A query search does paginate by default, as the results are likely to be smaller. Finally, if you want to create the repos that you have (the ones that don’t exist in the research software encyclopedia yet) just run create. The results are returned above, but they are also saved to the client.

scraper.create()

Bio.Tools

The bio.tools database is a rich source of scientific software, specifically for biotools. It serves an application pgogramming interface that we can query with the Research Software Encyclopedia, making it a good scraper. To scrape latest (does not have pagination), simply do:

$ rse scrape biotools

To do a dry run:

$ rse scrape --dry-run biotools
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/smajidian/phaseme
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/COVIDep/COVIDep
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/NIB-SI/DiNAR
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/reproducible-biomedical-modeling/Biosimulations
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/Brazelton-Lab/seq-annot
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/gevaertlab/BetaVAEImputation
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/herrsalmi/FConverter
INFO:rse.main.scrapers.biotools:Found repository: https://github.com/yanzhanglab/Graph2GO

You can also search for a term (also using --dry-run if desired:

$ rse scrape --dry-run biotools docker

The within python interaction is the same, except you need to select the biotools named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('biotools')
from rse.main.scrapers import BioToolsScraper
scraper = BioToolsScraper()

Hal Research Software Database

The Hal Research Software Database exposes a search API, and the client here tries to search for a subset of software that have github somewhere in the search (for latest). If you search for a term, although you will get more results, it’s less likely to find GitHub or GitLab.

$ rse scrape hal

To do a dry run:

$ rse scrape --dry-run hal
INFO:rse.main.scrapers.hal:Found repository: github.com/DreamCloud-Project/AMALTHEA-Microworkload-Generator
INFO:rse.main.scrapers.hal:Found repository: github.com/DreamCloud-Project/AMALTHEA-SimGrid
INFO:rse.main.scrapers.hal:Found repository: github.com/genotoul-bioinfo/dgenies
INFO:rse.main.scrapers.hal:Found repository: github.com/gijut/gnucash
INFO:rse.main.scrapers.hal:Found repository: github.com/DreamCloud-Project/McSim-Cycle-accurate-Xbar
INFO:rse.main.scrapers.hal:Found repository: github.com/DreamCloud-Project/McSim-TLM-NoC
INFO:rse.main.scrapers.hal:Found repository: github.com/DreamCloud-Project/McSim-Cycle-accurate-NoC
INFO:rse.main.scrapers.hal:Found repository: github.com/HEML/HEML
INFO:rse.main.scrapers.hal:Found repository: github.com/kkjawz/coref-ee
INFO:rse.main.scrapers.hal:Found repository: github.com/linbox-team/linbox
Found 10 results

You can also search for a term (also using --dry-run if desired:

$ rse scrape --dry-run hal docker

The within python interaction is the same, except you need to select the biotools named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('hal')
from rse.main.scrapers import HalScraper
scraper = HalScraper()

Research Software Dictionary (NL)

The Research Software Dictionary is a dictionary of software deployed by the Netherlands eScience Center that also provides a nice application programming interface (API) that we can parse.

$ rse scrape rsnl

To do a dry run:

$ rse scrape --dry-run rsnl
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/3D-e-Chem/knime-kripodb
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/3D-e-Chem/knime-pharmacophore
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/3D-e-Chem/knime-plants
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/3D-e-Chem/knime-python-node-archetype
...
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/iomega/spec2vec
INFO:rse.main.scrapers.rsnl:Found repository: https://github.com/NLESC-JCER/linux_actions_runner
Found 149 results

The within python interaction is the same, except you need to select the biotools named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('rsnl')
from rse.main.scrapers import RSNLScraper
scraper = RSNLScraper()

ROpenSci

The ROpenSci GitHub organization includes peer reviewed software that renders to https://docs.ropensci.org. We do this by way of parsing the ROpenSci GitHub repository (e.g., to get the latest or full listing) and also comparing this to the registry.json file. This means that we:

  1. Start with the GitHub repository listing, and skip over any repos not in the registry.
  2. We update the metadata with topics and description (if not defined) from the registry.
  3. In the case of a full parsing (not looking for latest) we add any names from the registry not seen in GitHub (e.g., repositories in other organizations)
  4. In the case of a “latest” parsing we do not consider this list.

This seems to do a fairly good job of capturing the bulk of ROpenSci repos! The “latest” scrape looks like this:

$ rse scrape ropensci

To do a dry run:

$ rse scrape --dry-run ropensci

The within python interaction is the same, except you need to select the ropensci named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('ropensci')
from rse.main.scrapers import ROpenSciScraper
scraper = ROpenSciScraper()

Molssi

The Molecular Sciences Software Institute maintains a paginated listing of software for computational chemistry, or more generally, molecular science at https://molssi.org/software-search/ that is accessible via the scaper here. Examples of command line usage include:

$ rse scrape molssi
$ rse scrape --dry-run molssi

The within python interaction is the same, except you need to select the molssi named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('molssi')

# or!
from rse.main.scrapers import MolssiScraper
scraper = MolssiScraperScraper()

Imperial College London Research Software Directory

You can browse the directory here! Note that if you want to use this as a Jekyll template (without relying on Alogia) @vsoch has prepared a template here.

$ rse scrape imperial
$ rse scrape --dry-run imperial

The within python interaction is the same, except you need to select the right named parser.

from rse.main.scrapers import get_named_scraper
scraper = get_named_scraper('imperial')

# or!
from rse.main.scrapers import ImperialCollegeLondonScraper
scraper = ImperialCollegeLondonScraper()

Astrophysics Source Code Library

You can browse the software in the library at https://ascl.net/code/all.

$ rse scrape ascl
$ rse scrape --dry-run ascl

And you can get a named scraper as shown in the other examples above.

BioGrids by Harvard Medical School

You can browse the software here. This is the first scraper that allows adding under the custom namespace, e.g., custom/biogrids.

$ rse scrape biogrids
$ rse scrape --dry-run biogrids

Debian Med

You can browse Debian Med categories and software here.

$ rse scrape debian
$ rse scrape --dry-run debian