SParse Lexical AnD Expansion Model for First Stage Ranking
This repository contains the code to perform training, indexing and retrieval for SPLADE models. It also includes everything needed to launch evaluation on the BEIR benchmark.
TL; DR SPLADE is a neural retrieval model which learns query/document sparse expansion via the BERT MLM head and sparse regularization. Sparse representations benefit from several advantages compared to dense approaches: efficient use of inverted index, explicit lexical match, interpretability... They also seem to be better at generalizing on out-of-domain data (cf BEIR benchmark).
- [v1]: SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking, Thibault Formal, Benjamin Piwowarski and Stéphane Clinchant. SIGIR21 short paper. link
By benefiting from recent advances in training neural retrievers, our v2 models rely on hard-negative mining, distillation and better Pre-trained Language Model initialization to further increase their effectiveness, on both in-domain (MS MARCO) and out-of-domain evaluation (BEIR benchmark).
- [v2]: SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval, Thibault Formal, Benjamin Piwowarski, Carlos Lassance, and Stéphane Clinchant. arxiv. link
- [v2bis]: From Distillation to Hard Negative Sampling: Making Sparse Neural IR Models More Effective, Thibault Formal, Carlos Lassance, Benjamin Piwowarski, and Stéphane Clinchant. SIGIR22 short paper. link
Weights for models trained under various settings can be found on Naver Labs Europe website, as well as Hugging Face. Please bear in mind that SPLADE is more a class of models rather than a model per se: depending on the regularization magnitude, we obtain very different models (from very sparse to models doing intense query/doc expansion) with different properties and performance.
splade: a spork that is sharp along one edge or both edges, enabling it to be used as a knife, a fork and a spoon.
In order to get the weights from this repo, be sure to have git lfs installed (not mandatory). We recommend to start
from a fresh environment, and install the packages from conda_splade_env.yml
.
conda create -n splade_env python=3.9
conda activate splade_env
conda env create -f conda_splade_env.yml
inference_splade.ipynb
allows you to load and perform inference with a trained model, in order to inspect the
predicted "bag-of-expanded-words". We provide weights for four main models:
model | MRR@10 (MS MARCO dev) |
---|---|
splade_max (v2) |
34.0 |
distilsplade_max (v2) |
36.8 |
naver/splade-cocondenser-selfdistil (v2bis, HF) |
37.6 |
naver/splade-cocondenser-ensembledistil (v2bis, HF) |
38.3 |
We also uploaded various models here. Feel free to try them out!
- This repository lets you either train (
train.py
), index (index.py
), retrieve (retrieve.py
) (or perform every step withall.py
) SPLADE models. - To manage experiments, we rely on hydra. Please refer to conf/README.md for a complete guide on how we configured experiments.
- To train models, we rely on MS MARCO data.
- We also further rely on distillation and hard negative mining, from available datasets (Margin MSE Distillation , Sentence Transformers Hard Negatives) or datasets we built ourselves (e.g. negatives mined from SPLADE).
- Most of the data formats are pretty standard; for validation, we rely on an (approximate) validation set, following a setting similar to TAS-B.
To simplify setting up, we made available all our data folders, which can
be downloaded here. This link includes queries,
documents and hard negative data, allowing for training under the EnsembleDistil
setting (see [v2 bis] paper). For
other settings (Simple
, DistilMSE
, SelfDistil
), you also have to download:
- (
Simple
) standard BM25 Triplets - (
DistilMSE
) "Vienna" triplets for MarginMSE distillation - (
SelfDistil
) triplets mined from SPLADE
After downloading, you can just untar in the root directory, and it will be placed in the right folder.
tar -xzvf file.tar.gz
In order to perform all steps (here on toy data, i.e. config_default.yaml
), go on the root directory and run:
conda activate splade_env
export PYTHONPATH=$PYTHONPATH:$(pwd)
export SPLADE_CONFIG_NAME="config_default.yaml"
python3 -m src.all \
config.checkpoint_dir=experiments/debug/checkpoint \
config.index_dir=experiments/debug/index \
config.out_dir=experiments/debug/out
We provide additional examples that can be plugged in the above code. See conf/README.md for details on how to change experiments settings.
- you can similarly run training
python3 -m src.train
(same for indexing or retrieval) - to create Anserini readable files (after training),
run
SPLADE_CONFIG_FULLPATH=/path/to/checkpoint/dir/config.yaml python3 -m src.create_anserini quantization_factor_document=100 quantization_factor_query=100
- config files for various settings (distillation etc.) are available in
/conf
. For instance, to run theSelfDistil
setting:- change to
SPLADE_CONFIG_NAME=config_splade _selfdistil.yaml
- to further change parameters (e.g. lambdas),
run:
python3 -m src.all config.regularizer.FLOPS.lambda_q=0.06 config.regularizer.FLOPS.lambda_d=0.02
- change to
We provide several base configurations which correspond to the settings in the [v2bis] paper. Please note that these are suited for our hardware setting, i.e. 4 GPUs Tesla V100 with 32GB memory. In order to train models with e.g. one GPU, you need to decrease the batch size for training and evaluation. Also note that, as the range for the loss might change with a different batch size, corresponding lambdas for regularization might need to be adapted.
Indexing (and retrieval) can be done either using our (numba-based) implementation of inverted index,
or Anserini. Let's perform these steps using an available model (naver/splade-cocondenser-ensembledistil
).
conda activate splade_env
export PYTHONPATH=$PYTHONPATH:$(pwd)
export SPLADE_CONFIG_FULLPATH="config_default.yaml"
python3 -m src.index \
init_dict.model_type_or_dir=naver/splade-cocondenser-ensembledistil \
config.pretrained_no_yamlconfig=true \
config.index_dir=experiments/pre-trained/index
python3 -m src.retrieve \
init_dict.model_type_or_dir=naver/splade-cocondenser-ensembledistil \
config.pretrained_no_yamlconfig=true \
config.index_dir=experiments/pre-trained/index \
config.out_dir=experiments/pre-trained/out
# pretrained_no_yamlconfig indicates that we solely rely on a HF-valid model path
- To change the data, simply override the hydra retrieve_evaluate package:
retrieve_evaluate=msmarco
You can similarly build the files that will be ingested by Anserini:
python3 -m src.create_anserini \
init_dict.model_type_or_dir=naver/splade-cocondenser-ensembledistil \
config.pretrained_no_yamlconfig=true \
config.index_dir=experiments/pre-trained/index \
quantization_factor_document=100 \
quantization_factor_query=100
It will create the json collection (docs_anserini.jsonl
) as well as the queries (queries_anserini.tsv
) that are
needed for Anserini. You then just need to follow the regression for
SPLADE here)
.
You can also run evaluation on BEIR, for instance:
conda activate splade_env
export PYTHONPATH=$PYTHONPATH:$(pwd)
export SPLADE_CONFIG_FULLPATH="/path/to/checkpoint/dir/config.yaml"
for dataset in arguana fiqa nfcorpus quora scidocs scifact trec-covid webis-touche2020 climate-fever dbpedia-entity fever hotpotqa nq
do
python3 -m src.beir_eval \
beir.dataset=$dataset \
beir.dataset_path=data/beir \
config.index_retrieve_batch_size=100
done
Please cite our work as:
- [v1] SIGIR21 short paper
@inbook{10.1145/3404835.3463098,
author = {Formal, Thibault and Piwowarski, Benjamin and Clinchant, St\'{e}phane},
title = {SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking},
year = {2021},
isbn = {9781450380379},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3404835.3463098},
booktitle = {Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval},
pages = {2288–2292},
numpages = {5}
}
- [v2] (arxiv)
@misc{https://doi.org/10.48550/arxiv.2109.10086,
doi = {10.48550/ARXIV.2109.10086},
url = {https://arxiv.org/abs/2109.10086},
author = {Formal, Thibault and Lassance, Carlos and Piwowarski, Benjamin and Clinchant, Stéphane},
keywords = {Information Retrieval (cs.IR), Artificial Intelligence (cs.AI), Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval},
publisher = {arXiv},
year = {2021},
copyright = {Creative Commons Attribution Non Commercial Share Alike 4.0 International}
}
- [v2bis] SIGIR22 short paper
TODO
SPLADE Copyright (c) 2021-present NAVER Corp.
SPLADE is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License. (see license)
You should have received a copy of the license along with this work. If not, see http://creativecommons.org/licenses/by-nc-sa/4.0/ .