File size: 798 Bytes
49a6ee1 19a28b1 49a6ee1 6dae94f 3ea99da 7f0aaff |
1 2 3 4 5 6 7 8 9 10 11 12 |
---
language: en
tags:
- allennlp
- question-answering
---
This is an implementation of the BiDAF model with ELMo embeddings. The basic layout is pretty simple: encode words as a combination of word embeddings and a character-level encoder, pass the word representations through a bi-LSTM/GRU, use a matrix of attentions to put question information into the passage word representations (this is the only part that is at all non-standard), pass this through another few layers of bi-LSTMs/GRUs, and do a softmax over span start and span end.
CAVEATS:
------
This model is based on ELMo. ELMo is not deterministic, meaning that you will see slight differences every time you run it. Also, ELMo likes to be warmed up, so we recommend processing dummy input before processing real workloads with it. |