Our BOW-BERT model is hosted on the huggingface model hub:
from transformers import AutoModelForSequenceClassification, AutoTokenizer
# load model
model = AutoModelForSequenceClassification.from_pretrained('dmrau/bow-bert')
# load tokenizer
tokenizer = AutoTokenizer.from_pretrained('bert-base-uncased')
# tokenize query and passage and concatenate them
inp = tokenizer(['this is a query','query a is this'], ['this is a passage', 'passage a is this'], return_tensors='pt')
# get estimated score
print('score', model(**inp).logits[:, 1])
### outputs identical scores for different
### word orders as the model is order invariant:
# scores: [-2.9463, -2.9463]
MSMARCO Testset 2020 on Nist Judgements:
NDCG@10: 65.79
MAP@1000: 42.64
RR@1000: 82.82
Recall@100: 68.21