mirror of https://github.com/microsoft/autogen.git
precommit: end-of-file-fixer (#929)
* precommit: end-of-file-fixer * exclude .gitignore * apply --------- Co-authored-by: Shaokun <shaokunzhang529@gmail.com>
This commit is contained in:
parent
a3e368d2ca
commit
2ff1035733
|
@ -159,4 +159,3 @@ automl.pkl
|
|||
|
||||
test/nlp/testtmp.py
|
||||
test/nlp/testtmpfl.py
|
||||
|
||||
|
|
|
@ -24,7 +24,9 @@ repos:
|
|||
- id: check-toml
|
||||
- id: check-json
|
||||
- id: check-byte-order-marker
|
||||
exclude: .gitignore
|
||||
- id: check-merge-conflict
|
||||
- id: detect-private-key
|
||||
- id: trailing-whitespace
|
||||
- id: end-of-file-fixer
|
||||
- id: no-commit-to-branch
|
|
@ -346,7 +346,10 @@ class AutoMLState:
|
|||
|
||||
@staticmethod
|
||||
def _compute_with_config_base(
|
||||
config_w_resource: dict, state: AutoMLState, estimator: str, is_report: bool = True
|
||||
config_w_resource: dict,
|
||||
state: AutoMLState,
|
||||
estimator: str,
|
||||
is_report: bool = True,
|
||||
) -> dict:
|
||||
if "FLAML_sample_size" in config_w_resource:
|
||||
sample_size = int(config_w_resource["FLAML_sample_size"])
|
||||
|
|
|
@ -23,5 +23,3 @@ Our findings on troubleshooting fine-tuning the Electra and RoBERTa model for th
|
|||
booktitle={ACL-IJCNLP},
|
||||
}
|
||||
```
|
||||
|
||||
|
||||
|
|
|
@ -2,4 +2,3 @@
|
|||
addopts = -m "not conda"
|
||||
markers =
|
||||
conda: test related to conda forge distribution
|
||||
|
||||
|
|
|
@ -131,4 +131,3 @@ module.exports = {
|
|||
],
|
||||
],
|
||||
};
|
||||
|
||||
|
|
|
@ -54,6 +54,3 @@
|
|||
]
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
|
Loading…
Reference in New Issue