| --- |
| language: multilingual |
| datasets: |
| - NQ |
| - Trivia |
| - SQuAD |
| - MLQA |
| - DRCD |
| --- |
| |
| # dpr-ctx_encoder-bert-base-multilingual |
| |
| ## Description |
| |
| Multilingual DPR Model base on bert-base-multilingual-cased. |
| [DPR model](https://arxiv.org/abs/2004.04906) |
| [DPR repo](https://github.com/facebookresearch/DPR) |
| |
| ## Data |
| 1. [NQ](https://github.com/facebookresearch/DPR/blob/master/data/download_data.py) |
| 2. [Trivia](https://github.com/facebookresearch/DPR/blob/master/data/download_data.py) |
| 3. [SQuAD](https://github.com/facebookresearch/DPR/blob/master/data/download_data.py) |
| 4. [DRCD*](https://github.com/DRCKnowledgeTeam/DRCD) |
| 5. [MLQA*](https://github.com/facebookresearch/MLQA) |
| |
| `question pairs for train`: 644,217 |
| `question pairs for dev`: 73,710 |
| |
| *DRCD and MLQA are converted using script from haystack [squad_to_dpr.py](https://github.com/deepset-ai/haystack/blob/master/haystack/retriever/squad_to_dpr.py) |
| |
| ## Training Script |
| I use the script from [haystack](https://colab.research.google.com/github/deepset-ai/haystack/blob/master/tutorials/Tutorial9_DPR_training.ipynb) |
| |
| ## Usage |
| |
| ```python |
| from transformers import DPRQuestionEncoder, DPRQuestionEncoderTokenizer |
| tokenizer = DPRQuestionEncoderTokenizer.from_pretrained('voidful/dpr-question_encoder-bert-base-multilingual') |
| model = DPRQuestionEncoder.from_pretrained('voidful/dpr-question_encoder-bert-base-multilingual') |
| input_ids = tokenizer("Hello, is my dog cute ?", return_tensors='pt')["input_ids"] |
| embeddings = model(input_ids).pooler_output |
| ``` |
| |
| Follow the tutorial from `haystack`: |
| [Better Retrievers via "Dense Passage Retrieval"](https://colab.research.google.com/github/deepset-ai/haystack/blob/master/tutorials/Tutorial6_Better_Retrieval_via_DPR.ipynb) |
| ``` |
| from haystack.retriever.dense import DensePassageRetriever |
| retriever = DensePassageRetriever(document_store=document_store, |
| query_embedding_model="voidful/dpr-question_encoder-bert-base-multilingual", |
| passage_embedding_model="voidful/dpr-ctx_encoder-bert-base-multilingual", |
| max_seq_len_query=64, |
| max_seq_len_passage=256, |
| batch_size=16, |
| use_gpu=True, |
| embed_title=True, |
| use_fast_tokenizers=True) |
| ``` |
| |