Go To:

Paper Title Paper Authors Table Of Contents Abstract References
Home
Report a problem with this paper

SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search

Authors

Abstract

With worldwide concerns surrounding the Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2), there is a rapidly growing body of scientific literature on the virus. Clinicians, researchers, and policy-makers need to be able to search these articles effectively. In this work, we present a zero-shot ranking algorithm that adapts to COVID-related scientific literature. Our approach filters training data from another collection down to medical-related queries, uses a neural re-ranking model pre-trained on scientific text (SciBERT), and filters the target document collection. This approach ranks top among zero-shot methods on the TREC COVID Round 1 leaderboard, and exhibits a P@5 of 0.80 and an nDCG@10 of 0.68 when evaluated on both Round 1 and 2 judgments. Despite not relying on TREC-COVID data, our method outperforms models that do. As one of the first search methods to thoroughly evaluate COVID-19 search, we hope that this serves as a strong baseline and helps in the global crisis.

1 Introduction

The emergence of the Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) prompted a worldwide research response. In the first 120 days of 2020, researchers published over 10,000 articles related to SARS-CoV-2 or COVID-19. Together with articles about similar viruses researched before 2020, the body of research approaches 60,000 articles. Such a large body of research results in a considerable burden for those seeking information about various facets of the virus, including researchers, clinicians, and policy-makers.

To help improve COVID-19 search, we introduce SLEDGE-Z: a simple yet effective zero-shot baseline for coronavirus Scientific knowLEDGE search. SLEDGE-Z adapts the successful BERTbased (Devlin et al., 2020 ) re-ranking model (Vanilla BERT, MacAvaney et al. (2019)) for COVID-19 search with three simple techniques. First, we propose a training data filtering technique to help the ranking model learn relevance signals typical in medical text. The training data we use comes entirely from another dataset (MS-MARCO, Campos et al. (2016) ), resulting in our model being zero-shot. Since MS-MARCO is a large collection of real user queries (over 800,000), it allows us to filter aggressively and still have adequate training data. Second, we replace the general contextualized language model BERT with one pretrained on scientific literature (SciBERT, Beltagy et al. (2019) ). This pre-training prepares the model for the type of language typically seen in scientific articles. Since the document collection (CORD-19, Wang et al. (2020) ) contains articles about prior viruses, we filter out articles published before 2020 to eliminate less pertinent articles. An overview of this process is shown in Figure 1 .

Figure 1: Overview of SLEDGE-Z.

We show that each of the techniques mentioned above positively impacts the ranking effectiveness of SLEDGE-Z through an ablation analysis. Our zero-shot approach performs comparably to (or outperforms) top-scoring submissions to the TREC-COVID document ranking shared task (Roberts et al., 2020) , a new testbed for evaluating of search methods for COVID-19. SLEDGE-Z tops the Round 1 leaderboard in the zero-shot setting, which is important in low-resource situations. Overall, our method establishes a strong performance for COVID-19 literature search. By releasing our models and code, we hope that it can help in the current global COVID-19 crisis. 1

2 Related Work

Ad-hoc document retrieval (of both scientific articles and general domain documents) has been long-studied (Lalmas and Tombros, 2007; Hersh and Voorhees, 2009; Lin, 2008; Medlar et al., 2016; Sorkhei et al., 2017; Huang et al., 2019; Hofstätter et al., 2020; Nogueira et al., 2020b) . Most recent work for scientific literature retrieval has focused on tasks such as collaborative filtering (Chen and Lee, 2018), citation recommendation (Nogueira et al., 2020a) , and clinical decision support (Soldaini et al., 2017) .

Pre-trained neural language models (such as BERT (Devlin et al., 2020) ) have recently shown to be effective when fine-tuned for ad-hoc ranking (Nogueira and Cho, 2019; Dai and Callan, 2019; MacAvaney et al., 2019) . These models also facilitate relevance signal transfer; Yilmaz et al. (2019) demonstrate that the relevance signals learned from BERT can transfer across collections (reducing the chance of overfitting a particular collection). Here, we use relevance signal transfer from an opendomain question answering dataset to the collection of COVID-19 scientific literature.

Others have investigated COVID-19 document ranking. Zhang et al. (2020) chronicled their efforts to build a search engine for COVID-19 articles, using a variety of available ranking techniques, such as T5 (Raffel et al., 2019) . In this work, we find that our approach outperforms this system in terms of ranking effectiveness. Contemporaneously with our work, Das et al. (2020) demonstrate how document clustering and summarization can be effective for COVID-19 retrieval. This paper extends our shared task submissions in Round 1 . We note that the TREC COVID task proceeded for a total of 5 rounds, with various techniques emerging, such as passage aggregation (Li et al., 2020; Nguyen et al., 2020) , and ensemble

3 Sledge-Z: Zero-Shot Covid-19 Search

To build a ranking model for COVID search, we modify the standard zero-shot Vanilla BERT document re-ranking pipeline (Yilmaz et al., 2019; MacAvaney et al., 2019) . We find that while these modifications are simple, they are effective for maximizing ranking performance. We note that this process neither requires COVID relevance training data nor involves a priori inspection of the queries and their characteristics. Thus, we consider our method zero-shot.

To train in a zero-shot setting, we employ a large dataset of general-domain natural language question and answer paragraphs: MS-MARCO (Campos et al., 2016) . However, naïve domain transfer is not optimal since most questions in the dataset are not medical-related, causing a domain mismatch between the training and evaluation data. To overcome this challenge, we apply a heuristic to filter the collection to only medical-related questions. The filter removes questions that do not contain terms appearing in the MedSyn (Yates and Goharian, 2013) , a lexicon of layperson and expert terminology for various medical conditions. We manually remove several common terms from the lexicon that commonly introduce queries that are not medical-related. For example, MedSyn includes the term gas (referring to the medical concept of flatulence in North American English), commonly also refers to gasoline or natural gas. See Appendix A.1 for a complete list of excluded MedSyn terms. Note that we made these decisions without considering COVID-19 specificallyonly a broad relation to the medical domain. MS-MARCO originally consists of 809K questions. After filtering, 79K of the original questions remain (9.7%). We refer to this subset of MS-MARCO as Med-MARCO. From a random sample of 100 queries from Med-MARCO, 78 were judged by the authors as medical-related, suggesting the filter has reasonable precision. Examples questions from this process include causes of peritoneal cancer prognosis and what is squalene anthrax sleep apnea. We make a list of the query IDs corresponding to Med-MARCO available, 2 as well as additional examples of filtered queries (see Appendix A.2).

Second, we replace the general-language BERT model with a variant tuned on scientific literature (including medical literature). Specifically, we use SciBERT (Beltagy et al., 2019) , which has an identical structure as BERT, but was trained on a multidomain corpus of scientific publications. It also uses a WordPiece lexicon based on the training data, allowing the model to better account for subwords commonly found in scientific text. During model training, we employ the pairwise cross-entropy loss function from Nogueira and Cho (2019) . Relevant and non-relevant documents are sampled in sequence from the official MS-MARCO training pair list (filtered down to Med-MARCO queries).

Third, we apply a filter to the document collection that removes any articles published before January 1, 2020. This filter aims to improve the retrieval system's precision by eliminating articles that may discuss other topics. The date was chosen because little was known about COVID-19 prior to 2020, and some documents do not include a full publication date (only a year), making this filter simple to apply. In real-life search engines, date filtering can often be applied at the discretion of the user.

4 Experimental Setup

We now explore the ranking effectiveness of our approach. We evaluate the performance of SLEDGE-Z using Round 1 and 2. At the time of writing, the only training data available for the task was the Round 1 data. of the TREC-COVID Information Retrieval Benchmark (Roberts et al., 2020 ). 3 TREC-COVID uses the CORD-19 document collection (Wang et al., 2020 ) (2020-05-01 version, 59,943 articles), with a set of 35 topics related to COVID-19. These topics include natural questions such as: what is the origin of COVID-19 and how does the coronavirus respond to changes in the weather. The top articles of participating systems in each round were judged by expert assessors, who rated each article as non-relevant (0), partiallyrelevant (1), or fully-relevant (2) to the topic. In total, 20,728 relevance judgments were collected (avg. 592 per topic), with 74% non-relevant, 12% partially relevant, and 14% fully-relevant. These rates remained nearly constant between rounds 1 and 2.

We use normalized Discounted Cumulative Gain with a cutoff of 10 (nDCG@10), Precision at 5 of partially and fully-relevant documents (P@5), and Precision at 5 of only fully relevant documents (P@5 (F)). Both nDCG@10 and P@5 are official task metrics; we include the P@5 filtered to only fully-relevance documents because it exposed some interesting trends in our analysis. We also report the percentage of the top 10 documents for each query that have relevance judgments (J@10). In an additional evaluation, we measure the performance using only judged documents to ensure that unjudged documents do not impact our findings. We used trec eval 4 for all metrics. These measures represent a precision-focused evaluation; since reranking methods like ours focus on improving precision, we leave recall-oriented evaluations to future work.

Our initial ranking is conducted using BM25 with default settings over the full document text to adhere to the zero-shot setting. Re-ranking is conducted over the abstracts only, avoiding the need to perform score aggregation (since BERT models are limited in the document length). We utilize only the natural-language question (ignoring the keyword query and extended narrative). We conduct an ablation that compares SLEDGE-Z to versions using BERT (instead of SciBERT), and the full MS-MARCO dataset (MSM) (rather than the Med-MARCO subset (MedM)). We compare with several baselines under the same evaluation settings.

-BM25: the initial BM25 ranking.

-ConvKNRM: The convolutional KNRM model (Dai et al., 2018) , trained on MS-MARCO data. -CEDR KNRM: The KNRM model, augmented with contextualized embeddings (MacAvaney et al., 2019), trained on MS-MARCO data. We use the bert-base-uncased model for the contextualized embeddings. -Seq2seq T5: The text-to-text-transformer (T5) model (Raffel et al., 2019) , tuned for ranking by predicting true or false as the next term in a sequence consisting of the query and document (Nogueira et al., 2020c) .

Including Unjudged

Judged Only

Model

Training nDCG@10 P@5 P@5 (F) J@10 nDCG@10 P@5 P@5 (F) -Fusion: a reciprocal rank fusion method (Cormack et al., 2009) of BM25 over the abstract, full text, and individual paragraphs. Fusion1 uses a concatenation of the keywords and question, and Fusion2 uses the entity extraction technique from the Round 1 udel submission. 5

Our work utilizes a variety of existing opensource tools: OpenNIR (MacAvaney, 2020), Anserini (Yang et al., 2017) , and the HuggingFace Transformers library (Wolf et al., 2019) . We utilize a held-out subset of 200 queries from the MS-MARCO training set as a validation set for the sole purpose of picking the optimal training epoch. Model hyper-parameters were chosen from values in prior work and can be found in Appendix A.4, along with information about the hardware used. The Vanilla BERT and SciBERT models take approximately 3 hours to train/validate, and inference on TREC-COVID takes approximately 15 minutes on modern GPUs. The BERT model has 157M parameters, and the SciBERT model has 158M parameters.

Training nDCG@10 P@5 P@5 (F) J@10 nDCG@10 P@5 P@5 (F) Table 2 : TREC COVID Round 1 and 2 comparison between SLEDGE-Z and other top official Round 2 submissions. We apply the date filter for a more complete comparison. Note that experimental differences exist between our system and these submissions, including the use of multiple topic fields and the utilization of Round 1 training data for training or tuning. The top result is marked in bold.

Table 2: TREC COVID Round 1 and 2 comparison between SLEDGE-Z and other top official Round 2 submissions. We apply the date filter for a more complete comparison. Note that experimental differences exist between our system and these submissions, including the use of multiple topic fields and the utilization of Round 1 training data for training or tuning. The top result is marked in bold.

System nDCG@10 P@5 P@5 (F) SLEDGE-Z (ours) 0.641 0.747 0.633 sab20.1.meta.docs 0.608 0.780 0.487 IRIT marked base 0.588 0.720 0.540 CSIROmedNIR 0.588 0.660 0.587 Table 3 : TREC-COVID Round 1 leaderboard (automatic systems). SLEDGE-Z outperforms the highestscoring run in terms of nDCG@10 and P@5 (F).

Table 3: TREC-COVID Round 1 leaderboard (automatic systems). SLEDGE-Z outperforms the highestscoring run in terms of nDCG@10 and P@5 (F).

et al., 2020) and SparseDenseSciBERT use relevant information from Round 1 as training data, and covidex.t5 uses combined keyword query and natural-language questions. Therefore, these performance metrics are not directly comparable to our zero-shot runs. Despite this, SLEDGE-Z still achieves competitive performance compared to these models. For instance, it consistently scores comparably or higher than covidex.t5 (includes a more powerful language model, a more effective initial ranking model, and multiple topic fields) and SparseDenseSciBert (which uses neural approaches for the initial ranking stage). Our method even performs comparably to the mpiid5.run3 model, which was trained directly on Round 1 judgments. Interestingly, we observe that our simple baseline approach of re-ranking using T5 strictly with the natural-language question against the paper title and abstract (Seq2seq T5 in Table 1 ) is more effective than the more involved approach employed by covidex.t5. When we apply the same date filtering to the official runs, we observe that the differences narrow. We also present SLEDGE-Z topping the Round 1 leaderboard in Table 3 . We observe again that our model excels at finding highly-relevant documents.

To gain a better understanding of the impact of filtering the document collection to only articles published on or after January 1, 2020, we first com-pare the performance of SLEDGE-Z with and without the filter. Disregarding unjudged documents, it has an nDCG@10 of 0.668 (−0.051), P@5 of 0.777 (−0.069) and P@5 (F) of 0.589 (−0.108). All these differences are statistically significant. By far the largest reduction is on fully-relevant P@5, meaning that it can be more difficult to find highly relevant documents when considering the full document collection. We observed similar trends for BM25, with and without the 2020 filter. These trends also align with observations we made from the judgments themselves; we find that only 16% of judged documents from prior to 2020 were considered relevant (with only 5% fully relevant). Meanwhile, 32% of judged documents after 2020 were considered relevant (19% fully relevant).

5 Results

Ranking effectiveness is presented in Table 1 . We first compare the ablations of our approach (top section). We note that SciBERT significantly (p < 0.05, paired t-test, Boneferroni correction) outperforms BM25 and BERT trained on MSM across all metrics. There is a less dramatic jump between BERT MSM and BERT MedM, demonstrating the importance of filtering the training data 5 https://github.com/castorini/anserini/ blob/master/docs/experiments-covid.md

Table 1: Ablation results and comparison of our approach and other zero-shot baselines on TREC-COVID Rounds 1 and 2. The top results are shown in bold. SciBERT with MedM (SLEDGE-Z) significantly outperforms values in the top (ablation) section marked with * (p < 0.05, paired t-test, Bonferroni correction).

properly. This is echoed between SciBERT MSM and SciBERT MedM, though the difference is only significant for P@5 when only considering the judged documents. These results demonstrate the importance of both pre-training on appropriate data and fine-tuning using a proper subset of the larger data. While both yield improvements (that can be additive), the pre-training objective appears to be more impactful, based on the overall better scores of SciBERT.

Compared to baseline systems (bottom section), we observe that SLEDGE-Z offers superior effectiveness. Specifically, we see that ConvKNRM, CEDR-KNRM, and Seq2seq T5 all improve upon the initial BM25 ranking. Training on Med-MARCO (rather than the full MS-MARCO) also improves each of the baselines, except, curiously, Seq2seq T5. This model may benefit from the larger amount of training data the full MS-MARCO dataset offers. Finally, both fusion methods outperform the base BM25 model. However, we note that these models utilize two fields available for each query: the keyword-based query and the full natural-language question text-a luxury not available in practical search environments. (Recall that SLEDGE-Z and the other baselines in Table 1 only use the natural-language query.)

We now compare our approach with the topperforming submissions to the TREC COVID shared task (many of which are not zero-shot methods). Full participating system descriptions are provided in Appendix A.3. We note that these experimental settings for these runs differ from our main experiments. For instance, mpiid5 run3 (Li Including Unjudged Judged Only

6 Conclusion

In this work, we present SLEDGE-Z, an adaptation of a neural ranking pipeline for COVID-19 scientific literature search. The approach is zeroshot and adapts to medical literature by filtering the training data, using a contextualized language model based trained on scientific text, and by filtering the document collection. The zero-shot setting is important because it suggests that the approach can be generally applied to similar problems-even when no training data are available (which can be expensive to collect). Through experiments and analysis on TREC-COVID, we find that each component of our approach is beneficial, and it outperforms or is comparable to approaches that are trained or tuned on TREC-COVID judgments. These observations underscore the importance of properly considering the domain when building medical search engines. We hope that techniques like SLEDGE-Z can help overcome the global COVID-19 crisis.

A.1 List Of Medsyn Exclusion Terms

The following terms were excluded from MedSyn when filtering MS-MARCO to reduce false positive matches: gas, card, bing, died, map, fall, falls. IRIT marked base: We use a BERT-base (12 layers, 768 hidden size) fine-tuned on Ms Marco passage set. We use a full ranking strategy with two stages: in the first stage, we use Anserini Bm25+ RM3 to retrieve top-1000 candidate documents for each topic using an index on the title+abstract of the CORD-19 documents, then we use the fine-tuned BERT to re-rank this list.

A.2 Med-Marco Examples

CSIROmedNIR: A neural index was built on the title, abstract fields of the COVID corpus alongside a traditional inverted index built on title, abstract and body text of the document. The neural index was built from the pooled classification token (1st token of the final BERT layer) using the covidbertnli model (https://huggingface.co/gsarti/covidbertnli) from the title, based off the sentence transformer (Reimers etal. Sentence-BERT, 2019). For the abstract,we took the Bag-of-Sentence approach where we averaged the individual sentence embeddings (sentence were segmented using segtok). All embeddings had a final dimension size of [1, 768] . We searched on the neural index using the query, narrative and question fields of the topics using the same embedding approach as with the document title embedding over the title and abstract neural index fields giving a total of 6 cosine similarity computations. We combine BM25 scores from traditional search over a combination of query,narrative and question fields over all document facets (body, title, abstract), giving a total of 9 different queryfacet combinations. We take the natural logarithm of the total BM25 score (to match the range of the cosine scores) which is then added the cosine scores: finalscore = log(sum of BM25 query-facet combs)+ cosine Scores Additionally, we filter the document by date. Documents created before December 31st 2019(before the first reported case) had their scores automatically set to zero.

Mpiid5.Run3:

We re-rank top-10000 documents returned by BM25 using the queries produced by Udel's method. For there-ranking method, we use the ELECTRA-Base model fine-tuned on the MSMARCO passage dataset. Themodel is later fine-tuned on the TREC COVID round 1 full-text collection.

SparseDenseSciBert: bm25+ann+scibert.0.33.teIn (ann-bm25 retrieval + scibert reranker): anserini TREC-COVID R2 Retrieval#8 + med-marco ANN + med-marco SciBERT with COVD Mask-Lm fine-tuning covidex.t5: Reciprocal rank fusion of two runs: Anserini r2.fusion1, reranked with medT5-3B; Anserini r2.fusion2, reranked with medT5-3B; Anserini fusion baselines for round 2: https://github.com/castorini/anserini/ blob/master/docs/expcovid.md medT5-3B: a T5-3B reranker fine-tuned on MS MARCO then fine-tuned (again) on MS MARCOmedical subset.

Code and models available at: https://github.com/ Georgetown-IR-Lab/covid-neural-ir.methods(Bendersky et al., 2020).

https://github.com/Georgetown-IR-Lab/ covid-neural-ir/blob/master/ med-msmarco-train.txt

Round 2 uses residual collection evaluation, meaning that all documents judged in Round 1 are disregarded. Although this is an important setting for building up a dataset and allows for approaches like manual relevance feedback, we feel that this setting does not mimic an actual search engine, especially in the zero-shot setting. Thus, we evaluate on the concatenation of Round 1 and 2 settings and mark the systems that use Round 1 judgments for training or tuning of their system.

https://github.com/usnistgov/trec_eval