Deep learning
Deep learning is an AI function and subset of machine learning, used for processing large amounts of complex data.
Here are 21,961 public repositories matching this topic...
Some lines in the code block of the keras docs is too long, the result of which is, there will be a horizonal scroll bar at the bottom of the code block. That is hard to read. The long lines should be rearranged to multiple short lines to improve readibility.
Example:
The docs for the SimpleRNN class (https://keras.io/layers/recurrent/#simplernn). The initializer of SimpleRNN has m
System information (version)
- OpenCV => master
Detailed description
This line of comment is inconsistent with the code:
https://github.com/opencv/opencv/blob/174b4ce29d8e1ddbd899095c4b9fb4443444af45/modules/imgcodecs/src/exif.hpp#L157
Constructor has been changed to use istream instead of filename in this PR:
opencv/opencv#8492
I think "outputs [-1]" and "outputs [0]" are equivalent (reversed) in this line of code, but the former (89%) works better than the latter (86%). Why?
🐛 Bug
To Reproduce
Run following from jupyter lab console
import torch
foo = torch.arange(5)
foo.as_strided((5,), (-1,), storage_offset=4)Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/daniil/.local/lib/python3.6/site-packages/torch/tensor.py", line 159, i
Line 1137 of the Caffe.Proto states "By default, SliceLayer concatenates blobs along the "channels" axis (1)."
Yet, the documentation on http://caffe.berkeleyvision.org/tutorial/layers/slice.html states, "The Slice layer is a utility layer that slices an input layer to multiple output layers along a given dimension (currently num or channel only) with given slice indices." which seems to be
-
Updated
Feb 22, 2020 - Python
can't find "from sklearn.cross_validation import train_test_split" in Latest version scikit-learn
Describe the bug
can't find "from sklearn.cross_validation import train_test_split" in Latest version scikit-learn
To Reproduce
Steps to reproduce the behavior:
- Day1
- Step 5: Splitting the datasets into training sets and Test sets
- Can't find "from sklearn.cross_validation import train_test_split" in Latest version scikit-learn**
**Desktop (please complete the following infor
I got this error:
Traceback (most recent call last):
File "c:\Users\jshat\Documents\Code\Machine Learning\Deep-Learning-Papers-Reading-Roadmap\download.py", line 88, in
readme_html = mistune.markdown(readme.read())
File "C:\Python37\lib\encodings\cp1252.py", line 23, in decode
return codecs.charmap_decode(input,self.errors,decoding_table)[0]
UnicodeDecodeError: 'charma
-
Updated
Feb 22, 2020 - Jupyter Notebook
-
Updated
Feb 22, 2020
This should really help to keep a track of papers read so far. I would love to fork the repo and keep on checking the boxes in my local fork.
For example: Have a look at this section. People fork this repo and check the boxes as they finish reading each section.
In Transformation Pipeline make class DataFrameSelector for custom transformation and call DataFrameSelector(num_attribs) it show
TypeError: object() takes no parameters
and same with CombinedAttributesAdder
i m using colab
from sklearn.base import BaseEstimator , TransformerMixin
class DataFrameSelector(BaseEstimator,TransformerMixin):
def _init_(self,attribute_names):
"Bokeh is a Python interactive visualization library that targets modern web browsers for presentation. Its goal is to provide elegant, concise construction of novel graphics in the style of D3.js, but also deliver this capability with high-performance interactivity over very large or streaming datasets. Bokeh can help anyone who would like to quickly and easi
-
Updated
Feb 22, 2020 - Jupyter Notebook
What's the ETA for updating the massively outdated documentation?
Please update all documents that are related building CNTK from source with latest CUDA dependencies that are indicated in CNTK.Common.props and CNTK.Cpp.props.
I tried to build from source, but it's a futile effort.
-
Updated
Feb 22, 2020 - Python
I am having difficulty in running this package as a Webservice. Would appreciate if we could provide any kind of documentation on implementing an API to get the keypoints from an image. Our aim is to able to deploy this API as an Azure Function and also know if it is feasible.
I was going though the existing enhancement issues again and though it'd be nice to collect ideas for spaCy plugins and related projects. There are always people in the community who are looking for new things to build, so here's some inspiration
If you have questions about the projects I suggested,
-
Updated
Feb 22, 2020 - Python
With the latest version of scipy.misc, scipy.misc.toimage is no longer available. To load and save an image as png we now have to use PIL, breaking tensorboard image summary.
Here is how I fixed the bug:
1./ At the end of main.py, log a uint8 image
logger.image_summary(tag, (images * 255).astype(np.uint8), step+1)
2./ In Logger class, package image as bytes with the PIL library (mode="L
From here:
A particularity of the SV2TTS framework is that all models can be trained
separately and on distinct datasets. For the encoder, one seeks to have a model
that is robust to noise and able to capture the many characteristics of the human
voice. Therefore, a large corpus of many different speakers wou
-
Updated
Feb 22, 2020 - Jupyter Notebook
-
Updated
Feb 22, 2020
-
Updated
Feb 22, 2020
-
Updated
Feb 22, 2020 - Python
It would be nice to have an API similar to strerror to get textual descriptions of error codes so applications can show something meaningful to users in error messages.
This was already implemented ad-hoc in the .NET bindings, see here: https://github.com/mozilla/DeepSpeech/blob/0b82c751db58d9d2d90e861f9af04e671fd2ab41/native_client/dotnet/DeepSpeechClien
-
Updated
Feb 22, 2020
-
Updated
Feb 21, 2020
-
Updated
Feb 21, 2020 - Lua
- Wikipedia
- Wikipedia
tf.functionmakes invalid assumptions about arguments that areMappinginstances. In general, there are no requirements forMappinginstances to have constructors that accept[(key, value)]initializers, as assumed here.This leads to cryptic exceptions when used with perfectly valid
Mappings