rigonsallauka
commited on
Commit
•
d28d47c
1
Parent(s):
2d1e6c3
Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
datasets:
|
4 |
-
-
|
5 |
language:
|
6 |
- es
|
7 |
metrics:
|
@@ -9,7 +9,6 @@ metrics:
|
|
9 |
- precision
|
10 |
- recall
|
11 |
- confusion_matrix
|
12 |
-
|
13 |
base_model:
|
14 |
- google-bert/bert-base-cased
|
15 |
pipeline_tag: token-classification
|
@@ -67,7 +66,7 @@ You can easily use this model with the Hugging Face `transformers` library. Here
|
|
67 |
```python
|
68 |
from transformers import AutoTokenizer, AutoModelForTokenClassification
|
69 |
|
70 |
-
model_name = "
|
71 |
|
72 |
# Load the tokenizer and model
|
73 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
datasets:
|
4 |
+
- HUMADEX/spanish_ner_dataset
|
5 |
language:
|
6 |
- es
|
7 |
metrics:
|
|
|
9 |
- precision
|
10 |
- recall
|
11 |
- confusion_matrix
|
|
|
12 |
base_model:
|
13 |
- google-bert/bert-base-cased
|
14 |
pipeline_tag: token-classification
|
|
|
66 |
```python
|
67 |
from transformers import AutoTokenizer, AutoModelForTokenClassification
|
68 |
|
69 |
+
model_name = "HUMADEX/spanish_medical_ner"
|
70 |
|
71 |
# Load the tokenizer and model
|
72 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|