AiLearning: 机器学习 - MachineLearning - ML、深度学习 - DeepLearning - DL、自然语言处理 NLP
-
Updated
Oct 28, 2019 - 13 commits
- Python
scikit-learn is a widely-used Python module for classic machine learning. It is built on top of SciPy.
AiLearning: 机器学习 - MachineLearning - ML、深度学习 - DeepLearning - DL、自然语言处理 NLP
Python Data Science Handbook: full text in Jupyter Notebooks
Data science Python notebooks: Deep learning (TensorFlow, Theano, Caffe, Keras), scikit-learn, Kaggle, big data (Spark, Hadoop MapReduce, HDFS), matplotlib, pandas, NumPy, SciPy, Python essentials, AWS, and various command lines.
A series of Jupyter notebooks that walk you through the fundamentals of Machine Learning and Deep Learning in python using Scikit-Learn and TensorFlow.
Your new Mentor for Data Science E-Learning.
The "Python Machine Learning (1st edition)" book code repository and info resource
Dive into Machine Learning with Python Jupyter notebook and scikit-learn!
Visualizer for neural network, deep learning and machine learning models
I've updated TPOT to version '0.10.2', and if I run the following code:
tpot = TPOTClassifier(generations=200, population_size=200, scoring='precision',
template='wrong template')
tpot.fit(X_train, y_train)I receive a RuntimeError:
`RuntimeError: There was an error in the TPOT optimization process. This could be because the data was not formatted
Open Machine Learning Course
Suppose I have a dask dataframe with lots of columns, all of them of float dtype. _meta_nonempty creates each column separately and then makes a dataframe out of it. This makes some operations surprisingly slow, such as loc. I see some issues with this:
_meta_nonempty can be optimized for some dataframes.loc shouldn't require _meta_nonempty in the first place. It only needs `sThe "Python Machine Learning (2nd edition)" book code repository and info resource
path_management fixture with pytest.tmpdir fixtureCurrently we define a custom fixture to clean up our serialization tests in entityset_tests.test_serialization. We can replace our custom fixture with pytest's tmpdir.
Hey all,
When running Auto-Sklearn in parallel with n_jobs and providing a True flag for delete_tmp_folder_after_terminate and delete_output_folder_after_terminate, the directories are not deleted in the end. This is not reflected in the documentation, where for both parameters the following is written:
delete_tmp_folder_after_terminate: string, optional (True)
remove tmp_folder,
PipelineAI: Real-Time Enterprise AI Platform
My blogs and code for machine learning. http://cnblogs.com/pinard
:book: [译] scikit-learn(sklearn) 中文文档
Jupyter notebooks from the scikit-learn video series
sklearn allows to pass a dictionary as scorer for multimetric scoring. For skorch, a user should pass multiple scoring callbacks instead.
Right now, if a user passes a dict, skorch will fail while trying to retrieve a name here:
Or here if a name was passed:
After the visualizer audit we believe that the quick methods are ready to be made prime time.
Quick methods must:
finalize() and not poof()For eac
Fit interpretable models. Explain blackbox machine learning.
A comprehensive list of Deep Learning / Artificial Intelligence and Machine Learning tutorials - rapidly expanding into areas of AI/Deep Learning / Machine Vision / NLP and industry specific areas such as Automotives, Retail, Pharma, Medicine, Healthcare by Tarry Singh until at-least 2020 until he finishes his Ph.D. (which might end up being inter-stellar cosmic networks! Who knows! 😀)
I think it could be useful, when one wants to plot only e.g. class 1, to have an option to produce consistent plots for both plot_cumulative_gain and plot_roc
At the moment, instead, only plot_roc supports such option.
Thanks a lot
A library for debugging/inspecting machine learning classifiers and explaining their predictions
As listed in https://docs.scipy.org/doc/scipy/reference/special.html#bessel-functions, we can implement the universal functions in the list first.
Python Cheat Sheet NumPy, Matplotlib
General Assembly's 2015 Data Science course in Washington, DC
I'm sorry if I missed this functionality, but CLI version hasn't it for sure (I saw the related code only in generate_code_examples.py). I guess it will be very useful to eliminate copy-paste phase, especially for large models.
Of course, piping is a solution, but not for development in Jupyter Notebook, for example.
Support for storing large tensor values in external files was introduced in #678, but AFAICT is undocumented.
This is a pretty important feature, functionally, but it's also important for end users who may not realise that they need to move around more than just the
*.onnxfile.I would suggest it should be documented in IR.md, and perhaps there are other locations from which it could be s