_config_info: create_time: '2023-03-14 23:35:21' use_default_base_config: true config_filepath: - /data/tzhu/MRC4IE/conf/mrc.yaml task_type: MrcQaTask task_name: RobertaBase_data20230314v2 comment: GlobalPointer with RoPE output_dir: outputs task_dir: outputs/RobertaBase_data20230314v2 dump_cache_dir: outputs/RobertaBase_data20230314v2/cache regenerate_cache: true data_dir: resources/Merged/20230314-v2 train_filepath: resources/Merged/20230314-v2/train.jsonl dev_filepath: resources/Merged/20230314-v2/dev.jsonl test_filepath: resources/Merged/20230314-v2/test.jsonl random_seed: 1227 num_epochs: 20 num_steps: -1 warmup_proportion: 0.1 epoch_patience: 10 step_patience: 5000 batch_size: 64 learning_rate: 2.0e-05 max_grad_norm: -1.0 skip_train: false debug_mode: false grad_accum_steps: 1 resumed_training_path: null step_eval_interval: 20000 epoch_eval_interval: 1 eval_on_data: - dev select_best_on_data: dev select_best_by_key: metric best_metric_field: micro.f1 save_every_ckpt: false save_best_ckpt: true final_eval_on_test: true main_process_logging: true max_seq_len: 512 plm_dir: hfl/chinese-roberta-wwm-ext train_batch_size: 16 eval_batch_size: 32 other_learning_rate: 2.0e-05 dropout: 0.3 biaffine_size: 512 include_package: - src.task use_default_base_config: true