|
_config_info: |
|
create_time: '2023-03-14 23:35:21' |
|
use_default_base_config: true |
|
config_filepath: |
|
- /data/tzhu/MRC4IE/conf/mrc.yaml |
|
task_type: MrcQaTask |
|
task_name: RobertaBase_data20230314v2 |
|
comment: GlobalPointer with RoPE |
|
output_dir: outputs |
|
task_dir: outputs/RobertaBase_data20230314v2 |
|
dump_cache_dir: outputs/RobertaBase_data20230314v2/cache |
|
regenerate_cache: true |
|
data_dir: resources/Merged/20230314-v2 |
|
train_filepath: resources/Merged/20230314-v2/train.jsonl |
|
dev_filepath: resources/Merged/20230314-v2/dev.jsonl |
|
test_filepath: resources/Merged/20230314-v2/test.jsonl |
|
random_seed: 1227 |
|
num_epochs: 20 |
|
num_steps: -1 |
|
warmup_proportion: 0.1 |
|
epoch_patience: 10 |
|
step_patience: 5000 |
|
batch_size: 64 |
|
learning_rate: 2.0e-05 |
|
max_grad_norm: -1.0 |
|
skip_train: false |
|
debug_mode: false |
|
grad_accum_steps: 1 |
|
resumed_training_path: null |
|
step_eval_interval: 20000 |
|
epoch_eval_interval: 1 |
|
eval_on_data: |
|
- dev |
|
select_best_on_data: dev |
|
select_best_by_key: metric |
|
best_metric_field: micro.f1 |
|
save_every_ckpt: false |
|
save_best_ckpt: true |
|
final_eval_on_test: true |
|
main_process_logging: true |
|
max_seq_len: 512 |
|
plm_dir: hfl/chinese-roberta-wwm-ext |
|
train_batch_size: 16 |
|
eval_batch_size: 32 |
|
other_learning_rate: 2.0e-05 |
|
dropout: 0.3 |
|
biaffine_size: 512 |
|
include_package: |
|
- src.task |
|
use_default_base_config: true |
|
|