Machine learning
Machine learning is the practice of teaching a computer to learn. The concept uses pattern recognition, as well as other forms of predictive algorithms, to make judgments on incoming data. This field is closely related to artificial intelligence and computational statistics.
Here are 41,326 public repositories matching this topic...
In the documentation it says:
Turns positive integers (indexes) into dense vectors of fixed size. eg. [[4], [20]] -> [[0.25, 0.1], [0.6, -0.2]]
Neither this explanation nor this example is very clear. I would suggest replacing this with
Turns positive integers (indexes) into dense vectors of fixed size. eg. [[4], [20]] -> [[0.25, 0.1], [0.6, -0.2]]
Describe the bug
Calling a pipeline with a nonparametric function causes an error since the function transform() is missing. The pipeline itself calls the function fit_transform() if it's present. For nonparametric functions (the most prominent being t-SNE) a regular transform() method does not exist since there is no projection or mapping that is learned. It could still be used f
trainable_variables = weights.values() + biases.values() doesn't work.
Also if I write trainable_variables = list(weights.values()) + list(biases.values()), I have to turn on tf.enable_eager_execution(), but the training result is wrong, accuracy is ar
Current Behavior:
The the wiki page APIExample, for the python example, the handle api is is run through the TessBaseAPIDelete funciton if the api failed to be initialized whereas for the C example below, this is not the case.
python:
rc = tesseract.TessBaseAPIInit3(api, TESSDATA_PREFIX, lang)
if (rc):
te-
Updated
Apr 9, 2020 - Python
-
Updated
Mar 5, 2020 - Python
From the test in #34944:
julia> REPLTests.fake_repl() do in, out, repl
repltask = @async begin
REPL.run_repl(repl)
end
write(in, "?;\n")
write(in, '\x04')
wait(repltask)
end
search:
-
Updated
Mar 30, 2020 - Jupyter Notebook
-
Updated
Feb 12, 2020
-
Updated
Dec 25, 2019 - TeX
-
Updated
Apr 14, 2020 - Jupyter Notebook
I'm not sure if XGBoost s model is well calibrated with softmax. It would be nice to have a doc with various experiments including random forest, dart etc.
Alexnet implementation in tensorflow has incomplete architecture where 2 convolution neural layers are missing. This issue is in reference to the python notebook mentioned below.
-
Updated
Apr 14, 2020 - Jupyter Notebook
-
Updated
Apr 11, 2020 - Python
What's the ETA for updating the massively outdated documentation?
Please update all documents that are related building CNTK from source with latest CUDA dependencies that are indicated in CNTK.Common.props and CNTK.Cpp.props.
I tried to build from source, but it's a futile effort.
Hi,
I thank you very much for your excellent work. However, I endure difficulties understanding your code ; my Java and Python training explain it a bit, but I think it would be great for everyone if it was documented. I mean a short explanation of each - complex - function, its role and its algorithm. I would be especially interested in a documentation of the tracker.
I thank you very much in a
In doc.pyx' s line 590:
if not self.is_parsed:
raise ValueError(Errors.E029)
I can still do a good job of chunking by tokenization and pos tagging only, without the full parse. Also in some languages parse isn't available. This will leave more flexibilities to users. I can comment out this in my copy of spacy, but when I update spacy to a new release, I have to chang
-
Updated
Jan 29, 2020 - Python
-
Updated
Feb 18, 2020 - Jupyter Notebook
-
Updated
Apr 14, 2020 - Python
-
Updated
Mar 20, 2020
-
Updated
Jun 12, 2017
-
Updated
Mar 28, 2020
- Wikipedia
- Wikipedia
Please make sure that this is a bug. As per our
GitHub Policy,
we only address code/doc bugs, performance issues, feature requests and
build/installation issues on GitHub. tag:bug_template
System information
example script provided in TensorFlow): Yes