Professional Documents
Culture Documents
S2ORC: The Semantic Scholar Open Research Corpus
S2ORC: The Semantic Scholar Open Research Corpus
Kyle Lo†∗ Lucy Lu Wang†∗ Mark Neumann† Rodney Kinney† Daniel S. Weld†‡
†
Allen Institute for Artificial Intelligence
‡
Paul G. Allen School of Computer Science & Engineering, University of Washington
{kylel, lucyw}@allenai.org
Abstract
We introduce S2ORC,1 a large corpus of
81.1M English-language academic papers
spanning many academic disciplines. The cor-
arXiv:1911.02782v3 [cs.CL] 7 Jul 2020
Table 1: A comparison of S2ORC with other publicly-available academic text corpora. Of the other corpora:
PubMed Central (OA) links to PubMed, which contains 30M papers at the time of writing. AAN links to the
ACL Anthology (which contained 25k papers at the time of dataset construction, and 54k papers at the time of
writing). Saier and Färber (2019) is derived from arXiv and links to MAG (which contained 213M papers and
other non-paper documents at the time of dataset construction, and 226M nodes at the time of writing). RefSeer
links to CiteSeerX (which contained 1M papers at the time of dataset construction, and 6M papers at the time of
writing). S2ORC contains three times more full text papers than PubMed Central (OA), the next largest corpus
with bibliometric enhancements, while covering a more diverse set of academic disciplines. Citations in S2ORC
are linked to the full set of S2ORC papers, 81.1M paper nodes derived from Semantic Scholar. In addition,
the L AT E X subset of S2ORC captures additional structure omitted by Saier and Färber (2019), who also parse
L AT E X sources from arXiv.
†
Saier and Färber (2020) is an update to this work which now includes full text. It is released concurrently with this work.
Yet, existing corpora are not without their limi- In this paper, we describe the construction of
tations. Some cover a small number of papers (e.g. S2ORC (§2). We provide summary statistics of
AAN), are domain-specific (e.g. AAN, PubMed the corpus (§3) and evaluate the data quality (§4).
Central, Saier and Färber (2019)), or may not pro- We then evaluate a BERT model pretrained on
vide usable full text (e.g. Saier and Färber (2019) S2ORC (§5), and discuss potential applications to
and RefSeer). To address these issues, we intro- a variety of NLP and analysis tasks over academic
duce S2ORC,7 the Semantic Scholar8 Open Re- text (§6). Finally, we compare S2ORC with other
search Corpus, a large publicly-available collec- publicly-available academic text corpora (§7).
tion of 81.1M academic papers covering dozens
of academic disciplines. Each paper is associated 2 Constructing the corpus
with metadata and abstracts aggregated from hun- S2ORC is constructed using data from the Se-
dreds of trusted sources such as academic publish- mantic Scholar literature corpus (Ammar et al.,
ers and literature archives like PubMed and arXiv. 2018). Papers in Semantic Scholar are derived
Notably, we release structured, machine- from numerous sources: obtained directly from
readable full text extracted from PDFs for 8.1M publishers, from resources such as MAG, from
papers which we’ve identified as having open ac- various archives such as arXiv or PubMed, or
cess status. S2ORC full text preserves meaningful crawled from the open Internet. Semantic Scholar
structure, e.g., paragraph breaks, section headers, clusters these papers based on title similarity and
inline citation mentions, references to tables and DOI overlap, resulting in an initial set of approxi-
figures, and resolved citation links to other papers. mately 200M paper clusters.
Additionally, we provide 1.5M full text L AT E X To construct S2ORC, we must overcome chal-
parses from which we have extracted, in addition lenges in (i) paper metadata aggregation, (ii) iden-
to citations and references, the source text of ta- tifying open access publications, and (iii) cluster-
bles and mathematical formulas. As shown in Ta- ing papers, in addition to identifying, extracting,
ble 1, S2ORC provides substantially more struc- and cleaning the full text and bibliometric annota-
tured full text papers and covers a more diverse set tions associated with each paper. The pipeline for
of academic disciplines than other resources. creating S2ORC is:
7
pronounced “stork” 1) Process PDFs and L AT E X sources to derive
8
The papers included in S2ORC are a curated subset of metadata, clean full text, inline citations and
the papers in the Semantic Scholar literature graph (Ammar references, and bibliography entries,
et al., 2018) that focuses only on English-language papers
with abstracts or full text available. See §2.5 for details on 2) Select the best metadata and full text parses
filtering through Semantic Scholar papers. for each paper cluster,
3) Filter paper clusters with insufficient meta- inline citations in papers with BRACKET-style ci-
data or content, and tations, and (ii) false negatives resulting from an
4) Resolve bibliography links between paper inability to expand bracket citation ranges (e.g.
clusters in the corpus. “[3]-[5]” should be expanded to “[3], [4], [5]” be-
Details for these steps are provided below. See fore linking). False positives are detected using
Appendix §A for definitions of terminology. The regular expressions and removed from G ROBID
output of this pipeline is visualized in Figure 1. output. Bracket citation ranges are manually ex-
panded and linked to their corresponding bibliog-
2.1 Processing PDFs raphy entries. The resulting parses are expressed
We process PDFs from the Semantic Scholar cor- in JSON format.12
pus using S CIENCE PARSE v3.0.09 and G ROBID
v0.5.510 (Lopez, 2009). Our processing pipeline 2.2 Processing L AT E X source
is described below. L AT E X document source is available for a major-
Selecting PDFs We remove PDFs which are less ity of arXiv submissions, and where available, are
likely to be academic papers. S CIENCE PARSE and used to construct a full text parse. We retrieve
G ROBID are not optimized for processing non- body text, section headers, figure/table captions,
paper academic documents such as dissertations, table representations, equations, and inline cita-
reports, slides, etc., and this filtering step is neces- tions and references directly from L AT E X source.
sary to increase output data quality. See Appendix Inspired by Saier and Färber (2019), we first con-
§B for filter details. There are around 31.3M PDFs vert L AT E X source into XML documents and then
associated with approximately 200M initial paper extract structured information from the XML.
clusters, and 30.5M PDFs are selected for process- Due to direct access to source, the accuracy of
ing based on these filtering criteria. citation span, reference, caption, section header,
and equation detection is near-perfect. We process
Extracting structured data from PDFs We use 1.5M papers from L AT E X source derived from
S CIENCE PARSE to extract title and authors from arXiv, all of which are included as part of S2ORC.
each PDF.11 We then use G ROBID to process each Surprisingly, due to the diversity of ways in which
PDF. From the XML output of G ROBID, we ex- authors define metadata in L AT E X, the quality
tract (i) metadata such as title, authors, and ab- of metadata extracted from L AT E X documents is
stract, (ii) paragraphs from the body text orga- worse than those extracted from PDF. Therefore,
nized under section headings, (iii) figure and ta- we do not use L AT E X-derived metadata for paper
ble captions, (iv) equations, table content, headers, clustering or metadata selection.
and footers, which we remove from the body text,
(v) inline citations in the abstract and body text, 2.3 Selecting canonical metadata
(vi) parsed bibliography entries with title, authors,
Canonical values for title, authors and other meta-
year, and venue identified, and (vi) links between
data fields are selected from among the papers in a
inline citation mentions and their corresponding
cluster. First, if a cluster contains multiple PDFs,
bibliography entries.
we select one to be canonical. This can occur, for
Postprocessing G ROBID output We postpro- example, in a cluster containing an arXiv preprint
cess G ROBID output using regular expressions to and its eventual camera-ready version. We pref-
classify the parenthetical citation style of a pa- erentially select PDFs from open access sources
per as BRACKET (e.g. [2]), NAME - YEAR (e.g. and break ties by prioritizing PDFs for which there
ABC, 2019), or OTHER (superscripts and other exist richer publisher-provided metadata (e.g. ab-
mixed styles). We focus on addressing two types stract, year, venue, DOI). If the selected PDF is
of common errors in G ROBID’s inline citation ex- associated with publisher-provided metadata, we
tractions: (i) false positives resulting from super- select those publisher-provided metadata fields to
scripts or equation references being recognized as be canonical.
9
https://github.com/allenai/ In cases where publisher-provided metadata
science-parse is incomplete, we use majority voting to select
10
https://github.com/kermitt2/grobid
11 12
Our evaluations suggest S CIENCE PARSE outperforms The S2ORC data format is described at https://
G ROBID for title and author extraction. github.com/allenai/s2orc
canonical metadata values. We break ties by min- do not provide significant value to our dataset in
imizing the total number of sources from which their current state. Although these papers that lack
we select metadata (e.g., if IEEE provides title, text may be useful as cite-able nodes in S2ORC,
authors and abstract, DBLP provides title and au- they are generally of lower quality and are filtered
thors, and arXiv provides title and abstract, we pri- out of the corpus to improve corpus quality.
oritize selecting IEEE over the union of DBLP and
arXiv). S2ORC metadata fields include title, au- Filter Number of papers
thor, year, venue, journal, abstract, and identifiers
No title 20k
(DOI, PubMed, PubMed Central (PMC), arXiv,
No authors 0.3M
and ACL Anthology).
< 100 chars of text 80.0M
In cases where the title and authors are not pro-
Not English 15.2M
vided by any publishers, we derive the values for
these fields from the parsed PDF, prioritizing S CI - Table 2: Post-processing data quality filters for papers
ENCE PARSE over G ROBID . We further comment
on paper clustering as it pertains to metadata se-
lection in Appendix §C. 2.6 Linking bibliographies to papers
2.4 Assembling the corpus Each bibliography entry in both G ROBID and L A -
We construct the final corpus by assembling clus- T E X parses are linked to the most similar papers
tered paper metadata with G ROBID and L AT E X in the corpus. For linking, we score each bibli-
parse objects. We associate the G ROBID parse ography entry and paper cluster pair using a sim-
with the S2ORC paper object if a valid G ROBID ilarity score computed between their titles. Each
parse is produced from the PDF, and the PDF is title is first normalized (i.e. white spaces stripped,
open access. Open access status is assigned if lower-cased, special characters removed) and rep-
a paper is derived from arXiv, ACL Anthology, resented by its character 3-grams. The similarity
PubMed Central (OA), and/or associated with an score Stitle is computed as the harmonic mean be-
open-access DOI in the Unpaywall database.13 If tween a Jaccard index and a containment metric:
the PDF is not open access, we only include the
2×J ×C
bibliography from the G ROBID parse in S2ORC. Stitle = (1)
J +C
If arXiv L AT E X source is available for the paper
cluster, we also associate the L AT E X parse with where the Jaccard index J and containment metric
the S2ORC paper object. C are computed from the n-grams of the two titles
N1 and N2 as:
2.5 Filtering paper clusters
We further filter paper clusters to remove papers |N1 ∩ N2 |
J=
with (i) no title, (ii) no authors, (iii) fewer than |N1 ∪ N2 |
100 characters of abstract and body text, and (iv)
where English is not the primary language. The |N1 ∩ N2 |
C=
first three filters remove papers that provide little min (|N1 |, |N2 |)
value for bibliometric-based or text-based analy- For each bibliography entry, the bibliography-
ses. The English language filter14 reduces G RO - paper pair with the highest similarity score above
BID parsing errors. All filters are applied in series.
0.8 is output as the correct link. Otherwise, the
Subsequently, 95.5M paper clusters are filtered bibliography entry remains unlinked. We perform
out based on the aforementioned criteria and re- an evaluation of linking performance in §4.
moved from the corpus. The distribution of fil-
tered papers is given in Table 2. We note that 3 The S2ORC dataset
a large number of paper clusters are filtered out;
80.0M of these filtered clusters have no associated The resulting corpus consists of 81.1M pa-
publisher-provided abstract or associated PDF and pers. Our publisher-provided abstract coverage is
13 90.4%, or 73.4M papers. Our PDF coverage is
Unpaywall 2019-04-19 data dump
14
We use the cld2 tool for language detection with a 35.6%, or 28.9M papers. These PDFs are pro-
threshold of 0.9 over the English language score. cessed using the pipeline discussed in §2.1. The
Total papers 81.1M
Papers w/ PDF 28.9M (35.6%)
Papers w/ bibliographies 27.6M (34.1%)
Papers w/ G ROBID full text 8.1M (10.0%)
Papers w/ LaTeX full text 1.5M (1.8%)
Papers w/ publisher abstract 73.4M (90.4%)
Papers w/ DOIs 52.2M (64.3%)
Papers w/ Pubmed IDs 21.5M (26.5%)
Papers w/ PMC IDs 4.7M (5.8%)
Papers w/ ArXiv IDs 1.7M (2.0%)
Papers w/ ACL IDs 42k (0.1%)
Table 5: S2ORC-S CI BERT test results are comparable with reported S CI BERT test results on the set of tasks
and datasets from Beltagy et al. (2019), to which we refer the reader for descriptions. Reported statistics are span-
level F1 for NER, token-level F1 for PICO, dependency parsing (DEP), and macro-F1 for relation (REL) and text
(CLS) classification. We report micro-F1 for ChemProt. All S2ORC-S CI BERT results are the mean ± standard
deviation of 5 runs with different random seeds. Beltagy et al. (2019) do not report standard deviation or number
of runs.