aari1995 commited on
Commit
73211b0
1 Parent(s): 4595e14

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -38,7 +38,7 @@ pipeline_tag: sentence-similarity
38
 
39
  The successor of [German_Semantic_STS_V2](https://huggingface.co/aari1995/German_Semantic_STS_V2) is here and comes with loads of cool new features! Feel free to provide feedback on the model and what you would like to see next.
40
 
41
- **Note:** To run this model properly, you need to set "trust_remote_code=True".
42
 
43
  ## Major updates and USPs:
44
 
@@ -51,8 +51,12 @@ The successor of [German_Semantic_STS_V2](https://huggingface.co/aari1995/German
51
  - **Pooling Function:** Moving away from mean pooling towards using the CLS token. Generally seems to learn better after the stage-2 pretraining and allows for more flexibility.
52
  - **License:** Apache 2.0
53
 
 
 
54
  ## Usage:
55
 
 
 
56
  ```python
57
  from sentence_transformers import SentenceTransformer
58
 
@@ -79,7 +83,6 @@ embeddings = model.encode(sentences, convert_to_tensor=True).half()
79
  # Get the similarity scores for the embeddings
80
  similarities = model.similarity(embeddings, embeddings)
81
 
82
-
83
  ```
84
 
85
  ### Full Model Architecture
@@ -101,10 +104,11 @@ Evaluation to come.
101
  **Q: Is this Model better than V2?**
102
 
103
  **A:** In terms of flexibility-definitely. In terms of data-yes as well, as it is more up-to-date. In terms of benchmark they differ, while V3 is better for longer texts, V2 works very well for shorter texts. Keeping in mind that many benchmarks also do not cover cultural knowledge too well.
 
104
 
105
  **Q: How does the model perform vs. multilingual models?**
106
 
107
- **A:** There are really great multilingual models that will be very useful for many use-cases. This model shines with its cultural knowledge and knowledge about German people and behaviour.
108
 
109
  **Q: What is the trade-off when reducing the embedding size?**
110
 
 
38
 
39
  The successor of [German_Semantic_STS_V2](https://huggingface.co/aari1995/German_Semantic_STS_V2) is here and comes with loads of cool new features! Feel free to provide feedback on the model and what you would like to see next.
40
 
41
+ **Note:** To run this model properly, see "Usage".
42
 
43
  ## Major updates and USPs:
44
 
 
51
  - **Pooling Function:** Moving away from mean pooling towards using the CLS token. Generally seems to learn better after the stage-2 pretraining and allows for more flexibility.
52
  - **License:** Apache 2.0
53
 
54
+ (If you are looking for even better performance on tasks, but with a German knowledge-cutoff around 2020, check out [German_Semantic_V3b](https://huggingface.co/aari1995/German_Semantic_V3))
55
+
56
  ## Usage:
57
 
58
+ This model has some build-in functionality that is rather hidden. To profit from it, use this code:
59
+
60
  ```python
61
  from sentence_transformers import SentenceTransformer
62
 
 
83
  # Get the similarity scores for the embeddings
84
  similarities = model.similarity(embeddings, embeddings)
85
 
 
86
  ```
87
 
88
  ### Full Model Architecture
 
104
  **Q: Is this Model better than V2?**
105
 
106
  **A:** In terms of flexibility-definitely. In terms of data-yes as well, as it is more up-to-date. In terms of benchmark they differ, while V3 is better for longer texts, V2 works very well for shorter texts. Keeping in mind that many benchmarks also do not cover cultural knowledge too well.
107
+ If you are fine with the model not knowing about developments after early 2020, I'd suggest you use [German_Semantic_V3b](https://huggingface.co/aari1995/German_Semantic_V3).
108
 
109
  **Q: How does the model perform vs. multilingual models?**
110
 
111
+ **A:** There are really great multilingual models that will be very useful for many use-cases. This model shines with its cultural knowledge and knowledge about German people and behaviour.
112
 
113
  **Q: What is the trade-off when reducing the embedding size?**
114