aari1995 commited on
Commit
ed9f58d
1 Parent(s): fd2763e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -35,13 +35,13 @@ pipeline_tag: sentence-similarity
35
  ---
36
 
37
  # German Semantic V3
38
- ## and [**German_Semantic_V3b**](https://huggingface.co/aari1995/German_Semantic_V3b)
39
 
40
  The successor ofs [German_Semantic_STS_V2](https://huggingface.co/aari1995/German_Semantic_STS_V2) are here and come with loads of cool new features! While V3 is really knowledge-heavy, V3b is more focused on performance. Feel free to provide feedback on the model and what you would like to see next.
41
 
42
  **Note:** To run this model properly, see "Usage".
43
 
44
- ## Major updates and USPs:
45
 
46
  - **Flexibility:** Trained with flexible sequence-length and embedding truncation, flexibility is a core feature of the model. Yet, smaller dimensions bring a minor trade-off in quality.
47
  - **Sequence length:** Embed up to 8192 tokens (16 times more than V2 and other models)
@@ -54,7 +54,7 @@ The successor ofs [German_Semantic_STS_V2](https://huggingface.co/aari1995/Germa
54
 
55
  (If you are looking for even better performance on tasks, but with a German knowledge-cutoff around 2020, check out [German_Semantic_V3b](https://huggingface.co/aari1995/German_Semantic_V3))
56
 
57
- ## Usage:
58
 
59
  This model has some build-in functionality that is rather hidden. To profit from it, use this code:
60
 
@@ -86,7 +86,7 @@ similarities = model.similarity(embeddings, embeddings)
86
 
87
  ```
88
 
89
- ### Full Model Architecture
90
 
91
  ```
92
  SentenceTransformer(
@@ -96,11 +96,11 @@ SentenceTransformer(
96
  ```
97
 
98
 
99
- ## Evaluation
100
 
101
  Evaluation to come.
102
 
103
- ## FAQ
104
 
105
  **Q: Is this Model better than V2?**
106
 
@@ -126,10 +126,10 @@ Also, V3 uses cls_pooling while V3buses mean_pooling.
126
 
127
  **A:** Broadly speaking, when going from 1024 to 512 dimensions, there is very little trade-off (1 percent). When going down to 64 dimensions, you may face a decrease of up to 3 percent.
128
 
129
- ## Up next:
130
  German_Semantic_V3_Instruct: Guiding your embeddings towards self-selected aspects
131
 
132
- ## Thank You and Credits
133
 
134
  - To [jinaAI](https://huggingface.co/jinaai) for their BERT implementation that is used, especially ALiBi
135
  - To [deepset](https://huggingface.co/deepset) for the gbert-large, which is a really great model
 
35
  ---
36
 
37
  # German Semantic V3
38
+ ### and [**German_Semantic_V3b**](https://huggingface.co/aari1995/German_Semantic_V3b)
39
 
40
  The successor ofs [German_Semantic_STS_V2](https://huggingface.co/aari1995/German_Semantic_STS_V2) are here and come with loads of cool new features! While V3 is really knowledge-heavy, V3b is more focused on performance. Feel free to provide feedback on the model and what you would like to see next.
41
 
42
  **Note:** To run this model properly, see "Usage".
43
 
44
+ # Major updates and USPs:
45
 
46
  - **Flexibility:** Trained with flexible sequence-length and embedding truncation, flexibility is a core feature of the model. Yet, smaller dimensions bring a minor trade-off in quality.
47
  - **Sequence length:** Embed up to 8192 tokens (16 times more than V2 and other models)
 
54
 
55
  (If you are looking for even better performance on tasks, but with a German knowledge-cutoff around 2020, check out [German_Semantic_V3b](https://huggingface.co/aari1995/German_Semantic_V3))
56
 
57
+ # Usage:
58
 
59
  This model has some build-in functionality that is rather hidden. To profit from it, use this code:
60
 
 
86
 
87
  ```
88
 
89
+ # Full Model Architecture
90
 
91
  ```
92
  SentenceTransformer(
 
96
  ```
97
 
98
 
99
+ # Evaluation
100
 
101
  Evaluation to come.
102
 
103
+ # FAQ
104
 
105
  **Q: Is this Model better than V2?**
106
 
 
126
 
127
  **A:** Broadly speaking, when going from 1024 to 512 dimensions, there is very little trade-off (1 percent). When going down to 64 dimensions, you may face a decrease of up to 3 percent.
128
 
129
+ # Up next:
130
  German_Semantic_V3_Instruct: Guiding your embeddings towards self-selected aspects
131
 
132
+ # Thank You and Credits
133
 
134
  - To [jinaAI](https://huggingface.co/jinaai) for their BERT implementation that is used, especially ALiBi
135
  - To [deepset](https://huggingface.co/deepset) for the gbert-large, which is a really great model