Skip to content
#

hyperparameter-tuning

Here are 463 public repositories matching this topic...

nni
pkubik
pkubik commented Mar 14, 2022

Describe the issue:
During computing Channel Dependencies reshape_break_channel_dependency does following code to ensure that the number of input channels equals the number of output channels:

in_shape = op_node.auxiliary['in_shape']
out_shape = op_node.auxiliary['out_shape']
in_channel = in_shape[1]
out_channel = out_shape[1]
return in_channel != out_channel

This is correct

bug help wanted good first issue model compression

Notes, programming assignments and quizzes from all courses within the Coursera Deep Learning specialization offered by deeplearning.ai: (i) Neural Networks and Deep Learning; (ii) Improving Deep Neural Networks: Hyperparameter tuning, Regularization and Optimization; (iii) Structuring Machine Learning Projects; (iv) Convolutional Neural Networks; (v) Sequence Models

  • Updated Feb 7, 2022
  • Jupyter Notebook
Neuraxle
evalml
chukarsten
chukarsten commented Feb 15, 2022

In #3324 , we had to mark some tests as expected to fail since XGBoost was throwing a FutureWarning. The warning has been addressed in XGBoost, so we're just waiting for the PR merged to be released. This issue is discussed in the #3275 issue.

evalml/tests/component_tests/test_xgboost_classifier.py needs to have the @pytest.mark.xfail removed f

testing good first issue tech debt
OCTIS

Improve this page

Add a description, image, and links to the hyperparameter-tuning topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the hyperparameter-tuning topic, visit your repo's landing page and select "manage topics."

Learn more