Natural language processing
Natural language processing (NLP) is a field of computer science that studies how computers and humans interact. In the 1950s, Alan Turing published an article that proposed a measure of intelligence, now called the Turing test. More modern techniques, such as deep learning, have produced results in the fields of language modeling, parsing, and natural-language tasks.
Here are 16,680 public repositories matching this topic...
-
Updated
Sep 7, 2021 - Python
-
Updated
Sep 11, 2021 - Python
-
Updated
Nov 24, 2021 - Python
-
Updated
Nov 24, 2021 - Python
-
Updated
Jun 12, 2017
-
Updated
Nov 24, 2021 - Python
In gensim/models/fasttext.py:
model = FastText(
vector_size=m.dim,
vector_size=m.dim,
window=m.ws,
window=m.ws,
epochs=m.epoch,
epochs=m.epoch,
negative=m.neg,
negative=m.neg,
# FIXME: these next 2 lines read in unsupported FB FT modes (loss=3 softmax or loss=4 onevsall,
# or model=3 supervi-
Updated
Nov 8, 2021
-
Updated
Nov 22, 2021
Describe the bug
I'm having major trouble with from_csv.
Context: I'm writing tutorial for build simple text search engine with Jina + Hub. I don't want to include a whole section of processing datasets, hence just passing a CSV into from_csv. I tried with meme dataset (converted tsv) before, and now using [superhero dataset](https://www.kaggle.com/jonathanbesomi/superheroes-nlp-datas
Motivated by huggingface/transformers#12789 in Transformers, one welcoming change would be replacing assertions with proper exceptions. The only type of assertions we should keep are those used as sanity checks.
Currently, there is a total of 87 files with the assert statements (located under datasets and src/datasets), so when working on this, to manage the PR s
-
Updated
Nov 24, 2021 - Python
Is your feature request related to a problem? Please describe.
I typically used compressed datasets (e.g. gzipped) to save disk space. This works fine with AllenNLP during training because I can write my dataset reader to load the compressed data. However, the predict command opens the file and reads lines for the Predictor. This fails when it tries to load data from my compressed files.
-
Updated
Jun 26, 2021 - Jupyter Notebook
-
Updated
Dec 22, 2020 - Python
-
Updated
Nov 22, 2021 - JavaScript
-
Updated
Jul 25, 2021 - Jupyter Notebook
-
Updated
Jul 1, 2021 - Python
-
Updated
Nov 24, 2021 - TypeScript
-
Updated
Nov 20, 2021 - Java
-
Updated
Jul 6, 2021
-
Updated
Nov 24, 2021 - Python
-
Updated
Nov 2, 2021 - Python
-
Updated
Oct 22, 2020
-
Updated
Nov 4, 2021 - Python
Created by Alan Turing
- Wikipedia
- Wikipedia
Currently, the
EncoderDecoderModelclass in PyTorch automatically creates thedecoder_input_idsbased on thelabelsprovided by the user (similar to how this is done for T5/BART). This should also be implemented forTFEncoderDecoderModel, because currently users should manually providedecoder_input_idsto the model.One can take a look at the TF implementation