hyperparameter-optimization
Here are 589 public repositories matching this topic...
Describe the issue:
During computing Channel Dependencies reshape_break_channel_dependency does following code to ensure that the number of input channels equals the number of output channels:
in_shape = op_node.auxiliary['in_shape']
out_shape = op_node.auxiliary['out_shape']
in_channel = in_shape[1]
out_channel = out_shape[1]
return in_channel != out_channel
This is correct
-
Updated
May 20, 2022 - Python
Expected behavior
GridSampler should stop the optimization when all grids are evaluated.
Environment
- Optuna version: 3.0.0b1.dev
- Python version: 3.8.6
- OS: macOS-10.16-x86_64-i386-64bit
- (Optional) Other libraries and their versions:
Error messages, stack traces, or logs
See steps to reproduce.Steps to reproduce
In the following code, optimize s
Scikit-learn provides multi-class options for area under curve: https://scikit-learn.org/stable/modules/generated/sklearn.metrics.roc_auc_score.html
We should provide the most common ones, such as the OVO Macro averaging used by Auto-Gluon.
Related: awslabs/autogluon#1479
Add a scikit-learn compatible API wrapper of TabularPredictor:
- TabularClassifier
- TabularRegressor
Required functionality (may need more than listed):
- init API
- fit API
- predict API
- works in sklearn pipelines
-
Updated
May 21, 2022 - Python
-
Updated
Jan 3, 2022
-
Updated
Feb 3, 2022 - Python
-
Updated
Nov 19, 2021 - Python
I trained models on Windows, then I tried to use them on Linux, however, I could not load them due to an incorrect path joining. During model loading, I got learner_path in the following format experiments_dir/model_1/100_LightGBM\\learner_fold_0.lightgbm. The last two slashes were incorrectly concatenated with the rest part of the path. In this regard, I would suggest adding something like `l
Discussed in microsoft/FLAML#543
Originally posted by scvail195 May 9, 2022
Call to resource.setrlimit(resource.RLIMIT_AS, (memory_limit, hard)) causes error
<img width="1399" alt="Screen Shot 2022-05-05 flaml crash" src="https://user-images.githubusercontent.com/90455225/167453259-0e30f323-0ae6-46ae-ab4d-2
-
Updated
May 21, 2022 - Python
-
Updated
Feb 10, 2021 - Python
-
Updated
Apr 23, 2022 - Python
-
Updated
Jun 6, 2018 - Python
-
Updated
Apr 24, 2022 - Jupyter Notebook
-
Updated
Feb 6, 2021 - Python
-
Updated
Apr 4, 2022 - Jupyter Notebook
-
Updated
Feb 27, 2022 - Python
-
Updated
Jun 19, 2021
-
Updated
Oct 14, 2021 - JavaScript
-
Updated
May 20, 2022 - Python
-
Updated
May 21, 2022 - Python
-
Updated
Jan 20, 2021 - Python
-
Updated
Aug 15, 2018 - Python
-
Updated
Apr 22, 2022 - Python
If enter_data() is called with the same train_path twice in a row and the data itself hasn't changed, a new Dataset does not need to be created.
We should add a column which stores some kind of hash of the actual data. When a Dataset would be created, if the metadata and data hash are exactly the same as an existing Dataset, nothing should be added to the ModelHub database and the existing
Describe the bug
Code could be more conform to pep8 and so forth.
Expected behavior
Less code st
Improve this page
Add a description, image, and links to the hyperparameter-optimization topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the hyperparameter-optimization topic, visit your repo's landing page and select "manage topics."
What happened + What you expected to happen
The autoscaler pushes some logs to the Ray driver. The logs are prefixed with
(scheduler)which is misleading.The prefix should be
(autoscaler).Versions / Dependencies
Ray master.
Reproduction script
Submit a Ray task or actor which triggers upscaling to trigger the logs.
Issue Severity
Low: It annoys or frustrates me.