Update README.md
Browse files
README.md
CHANGED
@@ -418,7 +418,7 @@ This is a benchmark of Korean embedding models.
|
|
418 |
|
419 |
## Bias, Risks and Limitations
|
420 |
|
421 |
-
- Since the evaluation results are different for each domain, it is necessary to compare and evaluate the model in your own domain. In the Miracl benchmark, the evaluation was conducted using the Korean Wikipedia as a corpus, and in this case, the cosine_ndcg@10 score dropped by 0.
|
422 |
- Also, since the miracl benchmark consists of a corpus of relatively short strings, while the Korean Embedding Benchmark consists of a corpus of longer strings, this model may be more advantageous if the length of the corpus you want to use is long.
|
423 |
|
424 |
|
|
|
418 |
|
419 |
## Bias, Risks and Limitations
|
420 |
|
421 |
+
- Since the evaluation results are different for each domain, it is necessary to compare and evaluate the model in your own domain. In the Miracl benchmark, the evaluation was conducted using the Korean Wikipedia as a corpus, and in this case, the cosine_ndcg@10 score dropped by 0.02 points after learning. However, in the Auto-RAG benchmark, which is a financial domain, the ndcg score increased by 0.09 when it was top 1. This model may be advantageous for use in a specific domain.
|
422 |
- Also, since the miracl benchmark consists of a corpus of relatively short strings, while the Korean Embedding Benchmark consists of a corpus of longer strings, this model may be more advantageous if the length of the corpus you want to use is long.
|
423 |
|
424 |
|