mirror of
https://github.com/microsoft/autogen.git
synced 2025-11-03 03:10:04 +00:00
* precommit: end-of-file-fixer * exclude .gitignore * apply --------- Co-authored-by: Shaokun <shaokunzhang529@gmail.com>
26 lines
1.3 KiB
Markdown
26 lines
1.3 KiB
Markdown
# AutoML for NLP
|
|
|
|
This directory contains utility functions used by AutoNLP. Currently we support four NLP tasks: sequence classification, sequence regression, multiple choice and summarization.
|
|
|
|
Please refer to this [link](https://microsoft.github.io/FLAML/docs/Examples/AutoML-NLP) for examples.
|
|
|
|
|
|
# Troubleshooting fine-tuning HPO for pre-trained language models
|
|
|
|
The frequent updates of transformers may lead to fluctuations in the results of tuning. To help users quickly troubleshoot the result of AutoNLP when a tuning failure occurs (e.g., failing to reproduce previous results), we have provided the following jupyter notebook:
|
|
|
|
* [Troubleshooting HPO for fine-tuning pre-trained language models](https://github.com/microsoft/FLAML/blob/main/notebook/research/acl2021.ipynb)
|
|
|
|
Our findings on troubleshooting fine-tuning the Electra and RoBERTa model for the GLUE dataset can be seen in the following paper published in ACL 2021:
|
|
|
|
* [An Empirical Study on Hyperparameter Optimization for Fine-Tuning Pre-trained Language Models](https://arxiv.org/abs/2106.09204). Xueqing Liu, Chi Wang. ACL-IJCNLP 2021.
|
|
|
|
```bibtex
|
|
@inproceedings{liu2021hpo,
|
|
title={An Empirical Study on Hyperparameter Optimization for Fine-Tuning Pre-trained Language Models},
|
|
author={Xueqing Liu and Chi Wang},
|
|
year={2021},
|
|
booktitle={ACL-IJCNLP},
|
|
}
|
|
```
|