Data Science
Data science is an inter-disciplinary field that uses scientific methods, processes, algorithms, and systems to extract knowledge from structured and unstructured data. Data scientists perform data analysis and preparation, and their findings inform high-level decisions in many organizations.
Here are 17,755 public repositories matching this topic...
Describe the issue linked to the documentation
The utils.gen_batches function is not referenced in the documentation (see the search result page), while for example utils.gen_even_slices is (see the search result page). Is this expected?
Suggest a potential alte
Screenshot
Without Maximize
With Maximize
Without Maximize (bar graph)
)
Travis is not going to automatically offer the free tier for all open source projects; We likely want o migrate away from travis.
Setting up github actions to replace travis would be a welcomed contribution.
In recent versions (can't say from exactly when), there seems to be an off-by-one error in dcc.DatePickerRange. I set max_date_allowed = datetime.today().date(), but in the calendar, yesterday is the maximum date allowed. I see it in my apps, and it is also present in the first example on the DatePickerRange documentation page.
E
NOTE: we'll need to verify that this is indeed what we want the behavior to be before doing any work on this.
Consider the following sequence of actions a user might take
- Populate a
config.tomlfile by runningstreamlit config show > ~/.streamlit/config.toml - In the file, change the
server.portconfig option to something other than 8501 (the default), let's say it's changed to 8502
Currently, we rely on AllGatherGrad to compute gather for GPUs.
TODO:
- [] Extend this class to support TPU
- [] Add tests
-
Updated
Mar 12, 2021 - Jupyter Notebook
Not a high-priority at all, but it'd be more sensible for such a tutorial/testing utility corpus to be implemented elsewhere - maybe under /test/ or some other data- or doc- related module – rather than in gensim.models.word2vec.
Originally posted by @gojomo in RaRe-Technologies/gensim#2939 (comment)
-
Updated
May 20, 2020
-
Updated
Mar 16, 2021
-
Updated
Oct 16, 2020 - Jupyter Notebook
-
Updated
Mar 15, 2021
While setting train_parameters to False very often we also may consider disabling dropout/batchnorm, in other words, to run the pretrained model in eval mode.
We've done a little modification to PretrainedTransformerEmbedder that allows providing whether the token embedder should be forced to eval mode during the training phase.
Do you this feature might be handy? Should I open a PR?
-
Updated
Mar 15, 2021
-
Updated
Feb 14, 2021 - JavaScript
-
Updated
Jan 25, 2021 - Python
-
Updated
Mar 15, 2021 - Python
I'm using mxnet to do some work, but there is nothing when I search the mxnet trial and example.
Current pytorch implementation ignores the argument split_f in the function train_batch_ch13 as shown below.
def train_batch_ch13(net, X, y, loss, trainer, devices):
if isinstance(X, list):
# Required for BERT Fine-tuning (to be covered later)
X = [x.to(devices[0]) for x in X]
else:
X = X.to(devices[0])
...Todo: Define the argument `
-
Updated
Mar 9, 2021
-
Updated
Feb 28, 2021 - Jupyter Notebook
-
Updated
Mar 12, 2021 - Python
- Wikipedia
- Wikipedia
(e.g. for links and images), because some of these examples are now being rendered in the docs.
Added by @fchollet in requests for contributions.