Natural language processing
Natural language processing (NLP) is a field of computer science that studies how computers and humans interact. In the 1950s, Alan Turing published an article that proposed a measure of intelligence, now called the Turing test. More modern techniques, such as deep learning, have produced results in the fields of language modeling, parsing, and natural-language tasks.
Here are 13,223 public repositories matching this topic...
-
Updated
Jan 20, 2021 - Python
-
Updated
Feb 25, 2021 - Python
-
Updated
Mar 5, 2021 - Python
-
Updated
Mar 5, 2021 - Python
-
Updated
Jun 12, 2017
Not a high-priority at all, but it'd be more sensible for such a tutorial/testing utility corpus to be implemented elsewhere - maybe under /test/ or some other data- or doc- related module – rather than in gensim.models.word2vec.
Originally posted by @gojomo in RaRe-Technologies/gensim#2939 (comment)
-
Updated
Mar 2, 2021
-
Updated
Feb 28, 2021
-
Updated
Mar 5, 2021 - Python
-
Updated
Mar 5, 2021 - Python
-
Updated
Dec 22, 2020 - Python
-
Updated
Mar 4, 2021 - JavaScript
While setting train_parameters to False very often we also may consider disabling dropout/batchnorm, in other words, to run the pretrained model in eval mode.
We've done a little modification to PretrainedTransformerEmbedder that allows providing whether the token embedder should be forced to eval mode during the training phase.
Do you this feature might be handy? Should I open a PR?
-
Updated
Mar 5, 2021 - TypeScript
-
Updated
Jan 1, 2021 - Python
-
Updated
Oct 20, 2020 - Jupyter Notebook
-
Updated
Apr 20, 2020 - Jupyter Notebook
-
Updated
Mar 3, 2021 - Java
-
Updated
Mar 5, 2021 - Python
-
Updated
Oct 22, 2020 - Python
-
Updated
Oct 22, 2020
-
Updated
Sep 23, 2020 - Jupyter Notebook
Hi I would like to propose a better implementation for 'test_indices':
We can remove the unneeded np.array casting:
Cleaner/New:
test_indices = list(set(range(len(texts))) - set(train_indices))
Old:
test_indices = np.array(list(set(range(len(texts))) - set(train_indices)))
-
Updated
Jan 22, 2021 - Python
Created by Alan Turing
- Wikipedia
- Wikipedia
Hi, I am interested in using the DeBERTa model that was recently implemented here and incorporating it into FARM so that it can also be used in open-domain QA settings through Haystack.
Just wondering why there's only a Slow Tokenizer implemented for DeBERTa and wondering if there are plans to create the Fast Tokeni