Data Science
Data science is an inter-disciplinary field that uses scientific methods, processes, algorithms, and systems to extract knowledge from structured and unstructured data. Data scientists perform data analysis and preparation, and their findings inform high-level decisions in many organizations.
Here are 17,738 public repositories matching this topic...
It appears that the docs for Logistic Regression differ based on solvers and penalties. The "penalty" parameter states that "The ‘newton-cg’, ‘sag’ and ‘lbfgs’ solvers support only l2 penalties," while the "solver" parameter states that "‘newton-cg’, ‘lbfgs’, ‘sag’ and ‘saga’ handle L2 or no penalty" (attaching some screenshots). This was actually a little unclear to me, as I wasn't sure if the n
Screenshot
Without Maximize
With Maximize
Without Maximize (bar graph)
)
Travis is not going to automatically offer the free tier for all open source projects; We likely want o migrate away from travis.
Setting up github actions to replace travis would be a welcomed contribution.
In recent versions (can't say from exactly when), there seems to be an off-by-one error in dcc.DatePickerRange. I set max_date_allowed = datetime.today().date(), but in the calendar, yesterday is the maximum date allowed. I see it in my apps, and it is also present in the first example on the DatePickerRange documentation page.
E
I am not sure if this is something that needs to be resolved simply in the documentation, that all types are required to match, or if it should/could be accommodated through the backend.
Steps to reproduce
Value Step mismatch
Code snippet:
stiffness = st.number_input(
'stiffness, k, kip/in',
value = 406.74,
step = 1
)
Error returned:
Stre
Currently, we rely on AllGatherGrad to compute gather for GPUs.
TODO:
- [] Extend this class to support TPU
- [] Add tests
-
Updated
Mar 12, 2021 - Jupyter Notebook
Not a high-priority at all, but it'd be more sensible for such a tutorial/testing utility corpus to be implemented elsewhere - maybe under /test/ or some other data- or doc- related module – rather than in gensim.models.word2vec.
Originally posted by @gojomo in RaRe-Technologies/gensim#2939 (comment)
-
Updated
May 20, 2020
-
Updated
Mar 12, 2021
-
Updated
Oct 16, 2020 - Jupyter Notebook
-
Updated
Jul 31, 2020
While setting train_parameters to False very often we also may consider disabling dropout/batchnorm, in other words, to run the pretrained model in eval mode.
We've done a little modification to PretrainedTransformerEmbedder that allows providing whether the token embedder should be forced to eval mode during the training phase.
Do you this feature might be handy? Should I open a PR?
-
Updated
Mar 11, 2021
-
Updated
Feb 14, 2021 - JavaScript
-
Updated
Jan 25, 2021 - Python
-
Updated
Mar 14, 2021 - Python
I'm using mxnet to do some work, but there is nothing when I search the mxnet trial and example.
Current pytorch implementation ignores the argument split_f in the function train_batch_ch13 as shown below.
def train_batch_ch13(net, X, y, loss, trainer, devices):
if isinstance(X, list):
# Required for BERT Fine-tuning (to be covered later)
X = [x.to(devices[0]) for x in X]
else:
X = X.to(devices[0])
...Todo: Define the argument `
-
Updated
Feb 28, 2021 - Jupyter Notebook
-
Updated
Mar 12, 2021 - Python
-
Updated
Mar 9, 2021
- Wikipedia
- Wikipedia
(e.g. for links and images), because some of these examples are now being rendered in the docs.
Added by @fchollet in requests for contributions.