Data Science
Data science is an inter-disciplinary field that uses scientific methods, processes, algorithms, and systems to extract knowledge from structured and unstructured data. Data scientists perform data analysis and preparation, and their findings inform high-level decisions in many organizations.
Here are 17,779 public repositories matching this topic...
It appears that the docs for Logistic Regression differ based on solvers and penalties. The "penalty" parameter states that "The ‘newton-cg’, ‘sag’ and ‘lbfgs’ solvers support only l2 penalties," while the "solver" parameter states that "‘newton-cg’, ‘lbfgs’, ‘sag’ and ‘saga’ handle L2 or no penalty" (attaching some screenshots). This was actually a little unclear to me, as I wasn't sure if the n
Screenshot
Without Maximize
With Maximize
Without Maximize (bar graph)
)
Travis is not going to automatically offer the free tier for all open source projects; We likely want o migrate away from travis.
Setting up github actions to replace travis would be a welcomed contribution.
In recent versions (can't say from exactly when), there seems to be an off-by-one error in dcc.DatePickerRange. I set max_date_allowed = datetime.today().date(), but in the calendar, yesterday is the maximum date allowed. I see it in my apps, and it is also present in the first example on the DatePickerRange documentation page.
E
NOTE: we'll need to verify that this is indeed what we want the behavior to be before doing any work on this.
Consider the following sequence of actions a user might take
- Populate a
config.tomlfile by runningstreamlit config show > ~/.streamlit/config.toml - In the file, change the
server.portconfig option to something other than 8501 (the default), let's say it's changed to 8502
Currently, we rely on AllGatherGrad to compute gather for GPUs.
TODO:
- [] Extend this class to support TPU
- [] Add tests
-
Updated
Mar 16, 2021 - Jupyter Notebook
Not a high-priority at all, but it'd be more sensible for such a tutorial/testing utility corpus to be implemented elsewhere - maybe under /test/ or some other data- or doc- related module – rather than in gensim.models.word2vec.
Originally posted by @gojomo in RaRe-Technologies/gensim#2939 (comment)
-
Updated
May 20, 2020
-
Updated
Mar 16, 2021
-
Updated
Oct 16, 2020 - Jupyter Notebook
-
Updated
Mar 15, 2021
While setting train_parameters to False very often we also may consider disabling dropout/batchnorm, in other words, to run the pretrained model in eval mode.
We've done a little modification to PretrainedTransformerEmbedder that allows providing whether the token embedder should be forced to eval mode during the training phase.
Do you this feature might be handy? Should I open a PR?
-
Updated
Mar 15, 2021
-
Updated
Feb 14, 2021 - JavaScript
-
Updated
Jan 25, 2021 - Python
-
Updated
Mar 17, 2021 - Python
I'm using mxnet to do some work, but there is nothing when I search the mxnet trial and example.
Current pytorch implementation ignores the argument split_f in the function train_batch_ch13 as shown below.
def train_batch_ch13(net, X, y, loss, trainer, devices):
if isinstance(X, list):
# Required for BERT Fine-tuning (to be covered later)
X = [x.to(devices[0]) for x in X]
else:
X = X.to(devices[0])
...Todo: Define the argument `
-
Updated
Mar 9, 2021
-
Updated
Feb 28, 2021 - Jupyter Notebook
-
Updated
Mar 12, 2021 - Python
- Wikipedia
- Wikipedia
(e.g. for links and images), because some of these examples are now being rendered in the docs.
Added by @fchollet in requests for contributions.