maximuspowers commited on
Commit
9ddd25e
·
verified ·
1 Parent(s): 329eb21

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -20,7 +20,7 @@ Classifies 512 chunks of a news article as "Likely" or "Unlikely" biased/disinfo
20
 
21
  Fine-tuned BERT ([bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased)) on the `headline`, `aritcle_text` and `text_label` fields in the [News Media Bias Plus Dataset](https://huggingface.co/datasets/vector-institute/newsmediabias-plus).
22
 
23
- **This model was trained with weighted sampling so that each batch contains 50% 'Likely' examples and 50% 'Unlikely' examples.** The same model trained without weighted sampling is here, and got slightly better taining eval metrics. However, this model preformed better when predictions were evaluated by gpt-4o as a judge.
24
 
25
  ### Metics
26
 
 
20
 
21
  Fine-tuned BERT ([bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased)) on the `headline`, `aritcle_text` and `text_label` fields in the [News Media Bias Plus Dataset](https://huggingface.co/datasets/vector-institute/newsmediabias-plus).
22
 
23
+ **This model was trained with weighted sampling so that each batch contains 50% 'Likely' examples and 50% 'Unlikely' examples.** The same model trained without weighted sampling is [here](https://huggingface.co/maximuspowers/nmbp-bert-full-articles), and got slightly better taining eval metrics. However, this model preformed better when predictions were evaluated by gpt-4o as a judge.
24
 
25
  ### Metics
26