Boolean query construction is often critical for medical systematic review literature search. To create an effective Boolean query, systematic review researchers typically spend weeks coming up with effective query terms and combinations. One challenge to creating an effective systematic review Boolean query is to select effective MeSH Terms in the query. In our previous work, we created neural MeSH term suggestion methods and compared them to state-of-the-art MeSH term suggestion methods. We found neural MeSH term suggestion methods to be highly effective. In this package, We implement ours and others MeSH term suggestion methods and that is aimed at researchers who want to further investigate, create or deploy such type of methods.
To install our package use:
git clone https://github.com/ielab/meshsuggestlib.git
cd meshsuggestlib
pip install .
Our package depends on Tevatron. However, we found we can not use Tevatron version on pip, please use the following command to install editable version:
git clone https://github.com/texttron/tevatron
cd tevatron
pip install .
Please setup your local enviroment and install all required package using:
pip install -r requirements.txt
Download our fine-tuned model from model_link
Then put model insiede library as:
model/checkpoint-80000
model/PubMed-w2v.bin
To use umls or metamap method for suggestion, it requires building of elastic server for both methods. For UMLS, please follow instruction on umls_link; for MetaMAP, follow instruction on metamap_link Remember to put metamap' public_mm on the lib directory.
List of pre-built method option include:
- Original
- ATM
- MetaMAP
- UMLS
- Atomic-BERT
- Semantic-BERT
- Fragment-BERT
List of pre-defined dataset option include:
- CLEF-2017
- CLEF-2018
- CLEF-2019-dta
- CLEF-2019-intervention
Running the following sample code can output result using CLEF-2017 dataset using Semantic-BERT
python -m meshsuggestlib
--output_dir model/ #Unused rightnow, but required in training-arg
--model_dir model/checkpoint-80000/ #Model checkpoint for inference
--method Semantic-BERT #Method Name
--dataset CLEF-2017 #Dataset Name, either pre-defined or new dataset(see research section for how to build new dataset)
--output_file result/out.tsv #Result file name
--email [email protected] #Email for pubmed E-utility submission
--interpolation_depth 20 #Intermediate depth for interpolation
--q_max_len 20 #MeSH Term length after tokenized
--p_max_len 20 #Query Length after tokenized
--date_file #Date restriction for each query
--mesh_file #MeSH Term file, see default mesh term file format
--depth 1 #Final retrieval Depth
python -m meshsuggestlib
--output_dir model/ #Unused rightnow, but required in training-arg
--evaluate_run #Boolean option for evaluate
--qrel_file data.qrels #Qrel file for evaluation
--output_file result/out.tsv #suggestion output file for evaluation
Preliminary Result to check:
Dataset | CLEF-2017 | CLEF-2018 | CLEF-2019-dta | CLEF-2019-intervention |
---|---|---|---|---|
Method\Metric | Precision/F_1/Recall | Precision/F_1/Recall | Precision/F_1/Recall | Precision/F_1/Recall |
Original | 0.0303/0.0323/0.7695 | 0.0226/0.0415/0.8629 | 0.0246/0.0453/0.8948 | 0.0166/0.0217/0.7450 |
ATM | 0.0225/0.0215/0.7109 | 0.0306/0.0535/0.8224 | 0.0111/0.0207/0.8936 | 0.0155/0.0181/0.7087 |
MetaMAP | 0.0323/0.0304/0.7487 | 0.0336/0.0590/0.8085 | 0.0137/0.0254/0.8774 | 0.0187/0.0211/0.6790 |
UMLS | 0.0325/0.0300/0.7379 | 0.0325/0.0573/0.7937 | 0.0133/0.0249/0.8598 | 0.0169/0.0186/0.6861 |
Atomic-BERT | 0.0252/0.0243/0.7778 | 0.0283/0.0479/0.8452 | 0.0096/0.0180/0.8850 | 0.0062/0.0111/0.7586 |
Semantic-BERT | 0.0255/0.0243/0.7785 | 0.0309/0.0526/0.8403 | 0.0108/0.0202/0.8810 | 0.0108/0.0181/0.7507 |
Fragment-BERT | 0.0343/0.0325/0.7415 | 0.0388/0.0690/0.8034 | 0.0235/0.0364/0.8765 | 0.0224/0.0276/0.7165 |
From the preliminary result it's clear that the best performance almost always in Neural Suggestion Methods.
To use new dataset in our library, build dataset structure as follows:
.
├── ...
├── New-DATASET # DATASET Folder Name
│ ├── TOPIC1 # Topic Name
│ │ │── 1 # Clause Number
│ │ │ │── clause_no_mesh # MeSH Term removed Clause (one line)
│ │ │ └── keywords # keyword in Clause (one keyword per line)
│ │ │── 2
│ │ └── 3
│ ├── TOPIC2
│ └── data.qrels # relevance judgement of query documents, format per line {topicid 0 docid 1}
...
Then to use new dataset, use --dataset DATASET_NAME during suggestion.
Two parameter to change during inference, includes:
--model_dir #Model_checkpoint
--tokenizer_name_or_path #Model_tokenizer_path
To use new functions, modify user_defined_method function in Suggestion.py
Then to use new Search Function, use --METHOD NEW during suggestion.
(Not working anymore): You can also try out our MeSH Term Suggestion Tool Server for different Suggestion methods As UQcloud is not very stable, alternatively, you can setup your own server locally following server repo at [MeSH_Suggester_Server] https://github.com/wshuai190/MeSH_Suggester_Server
MeSH Suggester has been intergrated to Tera, you can try it here.
MeSH Term Suggestion for Systematic Review Literature Search
If you use the Mesh Suggester system in your research, please cite the following paper:
@inproceedings{wang2023mesh,
title={Mesh Suggester: A Library and System for Mesh Term Suggestion for Systematic Review Boolean Query Construction},
author={Wang, Shuai and Li, Hang and Zuccon, Guido},
booktitle={Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining},
pages={1176--1179},
year={2023}
}
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License.