transformer
Here are 1,029 public repositories matching this topic...
-
Updated
Oct 20, 2020 - Jupyter Notebook
Bidirectional RNN
Is there a way to train a bidirectional RNN (like LSTM or GRU) on trax nowadays?
-
Updated
Feb 4, 2021 - Python
chooses 15% of token
From paper, it mentioned
Instead, the training data generator chooses 15% of tokens at random, e.g., in the sentence my
dog is hairy it chooses hairy.
It means that 15% of token will be choose for sure.
From https://github.com/codertimo/BERT-pytorch/blob/master/bert_pytorch/dataset/dataset.py#L68,
for every single token, it has 15% of chance that go though the followup procedure.
PositionalEmbedding
-
Updated
Mar 3, 2021 - Python
-
Updated
Nov 13, 2020 - Python
-
Updated
May 3, 2017 - Java
-
Updated
Feb 17, 2021 - Jupyter Notebook
-
Updated
Feb 20, 2021 - Jupyter Notebook
-
Updated
Mar 5, 2021 - JavaScript
-
Updated
Feb 7, 2019 - Python
-
Updated
Feb 10, 2021 - Python
-
Updated
Mar 10, 2021 - Python
-
Updated
Mar 5, 2021 - Java
-
Updated
Mar 10, 2021 - Python
-
Updated
Dec 19, 2020 - Python
-
Updated
Feb 27, 2021 - Python
-
Updated
Jul 26, 2019 - Python
Hi, I am so interesting in your project, and wonder if you need contributor and how could I make my own contribution?
-
Updated
Jan 23, 2021 - Jupyter Notebook
-
Updated
Dec 25, 2020 - Python
-
Updated
May 12, 2020 - Java
-
Updated
Jan 21, 2021 - Python
-
Updated
Nov 15, 2019 - Python
-
Updated
Mar 9, 2021 - Python
-
Updated
Mar 10, 2021 - Python
-
Updated
Mar 7, 2021 - TypeScript
-
Updated
Jun 1, 2020
Improve this page
Add a description, image, and links to the transformer topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the transformer topic, visit your repo's landing page and select "manage topics."
Hi, I am interested in using the DeBERTa model that was recently implemented here and incorporating it into FARM so that it can also be used in open-domain QA settings through Haystack.
Just wondering why there's only a Slow Tokenizer implemented for DeBERTa and wondering if there are plans to create the Fast Tokeni