* added 'forecast' task with estimators ['fbprophet', 'arima', 'sarimax']
* update setup.py
* add TimeSeriesSplit to 'regression' and 'classification' task
* add 'time' split_type for 'classification' and 'regression' task
Signed-off-by: Kevin Chen <chenkevin.8787@gmail.com>
* feature importance
* variable name
* Update test/test_split.py
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Update test/test_forecast.py
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* prophet installation fail in windows
* upload flaml_forecast.ipynb
Signed-off-by: Kevin Chen <chenkevin.8787@gmail.com>
* remove extra comma
* exclusive bound
* log file name
* add cost to space
* dataset_format
* add load_openml_dataset test
* docstr
* revise test format
* simplify restore
* order categories
* openml server exception in test
* process space
* add warning
* log format
* reduce n_cpu
* nested space
* hierarchical search space for CFO
* non hierarchical for bs
* unflatten hierarchical config
* connection error
* random sample
* config signature
* check ray version
* preprocess numpy array
* catboost preprocess
* time budget
* seed, verbose, hpo_method
* test cfocat
* shallow copy in flatten_dict
prevent lgbm model duplication
* match estimator name
* quantize and log
* test qloguniform and qrandint
* test qlograndint
* thread.running
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
Co-authored-by: Qingyun Wu <qingyunwu@Qingyuns-MacBook-Pro-2.local>
* add starting point in fit
* add estimator best config
* add test
* add doc string
* when there are multiple points_to_evaluate in CFO, use the best one to start local search; after that use low cost partial config as the start point; then, remove the points whose performance is worse than the converged, and start local search from the remaining ones ordered by their performance.
Co-authored-by: Qingyun Wu <qingyunwu@Qingyuns-MacBook-Pro-2.local>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* pickle the AutoML object
* get best model per estimator
* test deberta
* stateless API
* pickle the AutoML object
* get best model per estimator
* test deberta
* stateless API
* prevent divide by zero
* test roberta
* BlendSearchTuner
* sync
* version number
* update gitignore
* delta time
* reindex columns when dropping int-indexed columns
* add seed
* add seed in Args
* merge
* stabilize SearchThread speed
* add seed
* fix import
* use except
* add restore test for CFO
* remove test_restore
* remove inspect
* remove print
* change to SearchThread._esp
* add _eps lower bound
* _eps in SearchThread
* add test_restore
* 1<<32
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
Co-authored-by: Qingyun Wu <qiw@microsoft.com>
* pickle the AutoML object
* get best model per estimator
* test deberta
* stateless API
* pickle the AutoML object
* get best model per estimator
* test deberta
* stateless API
* prevent divide by zero
* test roberta
* BlendSearchTuner
* sync
* version number
* update gitignore
* delta time
* reindex columns when dropping int-indexed columns
* add seed
* add seed in Args
* merge
* init upload of ChaCha
* remove redundancy
* add back catboost
* improve AutoVW API
* set min_resource_lease in VWOnlineTrial
* docstr
* rename
* docstr
* add docstr
* improve API and documentation
* fix name
* docstr
* naming
* remove max_resource in scheduler
* add TODO in flow2
* remove redundancy in rearcher
* add input type
* adapt code from ray.tune
* move files
* naming
* documentation
* fix import error
* fix format issues
* remove cb in worse than test
* improve _generate_all_comb
* remove ray tune
* naming
* VowpalWabbitTrial
* import error
* import error
* merge test code
* scheduler import
* fix import
* remove
* import, minor bug and version
* Float or Categorical
* fix default
* add test_autovw.py
* add vowpalwabbit and openml
* lint
* reorg
* lint
* indent
* add autovw notebook
* update notebook
* update log msg and autovw notebook
* update autovw notebook
* update autovw notebook
* add available strings for model_select_policy
* string for metric
* Update vw format in flaml/onlineml/trial.py
Co-authored-by: olgavrou <olgavrou@gmail.com>
* make init_config optional
* add _setup_trial_runner and update notebook
* space
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
Co-authored-by: Qingyun Wu <qiw@microsoft.com>
Co-authored-by: olgavrou <olgavrou@gmail.com>
* datetime feature engineering added.
* check if datetime in columns moved after drop check. Check if the new columns do not already exist.
* check the drop condition before to add new_column. In transform, check directly if new columns are present in num_column.
* check if new_column is in X.columns.
* fixed lint issue. update version to 0.4.1.
* add customized lgbm learner
* add comments
* fix format issue
* format
* OpenMLError
* add test
* add notebook
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* pickle the AutoML object
* get best model per estimator
* test deberta
* stateless API
* Add Gitter badge (#41)
* prevent divide by zero
* test roberta
* BlendSearchTuner
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>
Co-authored-by: The Gitter Badger <badger@gitter.im>
* xgboost notebook
* finetuning notebook
* finetuning test
* experimental nni support
* support nested search space
* log file name
* record training_iteration
* eps
* reset times
* std set to default step size if 0
* v0.2.2
separate the HPO part into the module flaml.tune
enhanced implementation of FLOW^2, CFO and BlendSearch
support parallel tuning using ray tune
add support for sample_weight and generic fit arguments
enable mlflow logging
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>
Co-authored-by: qingyun-wu <qw2ky@virginia.edu>
* set default logging level to INFO
* remove unnecessary import
* API future compatibility
* add test for customized learner
* test dependency
Co-authored-by: Chi Wang (MSR) <chiw@microsoft.com>