Commit
·
91d8b3e
1
Parent(s):
5ab02d7
modify readme
Browse files
README.md
CHANGED
|
@@ -14,7 +14,9 @@ pipeline_tag: sentence-similarity
|
|
| 14 |
|
| 15 |
## Description
|
| 16 |
|
| 17 |
-
This is a fine-tuned version of [MiniCPM-2B-dpo-bf16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16) to perform Text Embedding tasks. The model is fine-tuned using the Contrastive Fine-tuning and LoRA technique on NLI datasets.
|
|
|
|
|
|
|
| 18 |
|
| 19 |
## Base Model
|
| 20 |
|
|
@@ -88,7 +90,7 @@ print(encoded_sentences)
|
|
| 88 |
|
| 89 |
## Training Details
|
| 90 |
|
| 91 |
-
⚠️
|
| 92 |
|
| 93 |
| **Training Details** | **Value** |
|
| 94 |
|-------------------------|-------------------|
|
|
|
|
| 14 |
|
| 15 |
## Description
|
| 16 |
|
| 17 |
+
This is a fine-tuned version of [MiniCPM-2B-dpo-bf16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16) to perform Text Embedding tasks. The model is fine-tuned using the Contrastive Fine-tuning and LoRA technique on NLI datasets.
|
| 18 |
+
|
| 19 |
+
⚠️ The training process ignores hard-negative samples and treat other in-batch samples + their entailments as in-batch negatives. ⚠️ If you want to see the version utilizing hard-negative examples in the training process, please refer [here](https://huggingface.co/trapoom555/MiniCPM-2B-Text-Embedding-cft)
|
| 20 |
|
| 21 |
## Base Model
|
| 22 |
|
|
|
|
| 90 |
|
| 91 |
## Training Details
|
| 92 |
|
| 93 |
+
⚠️ The training process ignores hard-negative samples and treat other in-batch samples + their entailments as in-batch negatives. ⚠️
|
| 94 |
|
| 95 |
| **Training Details** | **Value** |
|
| 96 |
|-------------------------|-------------------|
|