Natural language processing
Natural language processing (NLP) is a field of computer science that studies how computers and humans interact. In the 1950s, Alan Turing published an article that proposed a measure of intelligence, now called the Turing test. More modern techniques, such as deep learning, have produced results in the fields of language modeling, parsing, and natural-language tasks.
Here are 18,850 public repositories matching this topic...
-
Updated
May 9, 2022 - Python
-
Updated
May 8, 2022 - Python
-
Updated
May 20, 2022 - Python
-
Updated
Jun 12, 2017
-
Updated
May 20, 2022 - Python
Describe the bug
Streaming Datasets can't be pickled, so any interaction between them and multiprocessing results in a crash.
Steps to reproduce the bug
import transformers
from transformers import Trainer, AutoModelForCausalLM, TrainingArguments
import datasets
ds = datasets.load_dataset('oscar', "unshuffled_deduplicated_en", split='train', streaming=True).with_format("-
Updated
May 19, 2022
In gensim/models/fasttext.py:
model = FastText(
vector_size=m.dim,
vector_size=m.dim,
window=m.ws,
window=m.ws,
epochs=m.epoch,
epochs=m.epoch,
negative=m.neg,
negative=m.neg,
# FIXME: these next 2 lines read in unsupported FB FT modes (loss=3 softmax or loss=4 onevsall,
# or model=3 supervi-
Updated
Apr 1, 2022 - Jupyter Notebook
-
Updated
May 22, 2022 - Python
Is your feature request related to a problem? Please describe.
I typically used compressed datasets (e.g. gzipped) to save disk space. This works fine with AllenNLP during training because I can write my dataset reader to load the compressed data. However, the predict command opens the file and reads lines for the Predictor. This fails when it tries to load data from my compressed files.
Checking the Python files in NLTK with "python -m doctest" reveals that many tests are failing. In many cases, the failures are just cosmetic discrepancies between the expected and the actual output, such as missing a blank line, or unescaped linebreaks. Other cases may be real bugs.
If these failures could be avoided, it would become possible to improve CI by running "python -m doctest" each t
-
Updated
Jul 25, 2021 - Jupyter Notebook
-
Updated
May 17, 2022 - JavaScript
-
Updated
Dec 22, 2020 - Python
-
Updated
May 12, 2022 - TypeScript
-
Updated
May 19, 2022 - Python
-
Updated
Jul 6, 2021
-
Updated
May 6, 2022 - Java
-
Updated
May 13, 2022 - Python
-
Updated
Nov 2, 2021 - Python
-
Updated
Oct 22, 2020
-
Updated
Mar 30, 2022 - Python
-
Updated
Apr 3, 2022 - Python
-
Updated
Sep 23, 2020 - Jupyter Notebook
Created by Alan Turing
- Wikipedia
- Wikipedia
Model description
I would like to add a new model:
Proposed in the paper: UNETR: Transformers for 3D Medical Image Segmentation
UNEt TRansformers (UNETR) utilize a transformer as the encoder to learn sequence representations of the input volume and effectively capture the global multi-scale information, while also following the successful "U-shaped"