asdc commited on
Commit
c771149
·
verified ·
1 Parent(s): 22ce76e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -7
README.md CHANGED
@@ -4,14 +4,18 @@ tags:
4
  model-index:
5
  - name: XLM_normalization_BEST_MODEL
6
  results: []
 
 
 
 
 
 
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
11
-
12
  # XLM_normalization_BEST_MODEL
13
 
14
- This model was trained from scratch on the None dataset.
 
15
 
16
  ## Model description
17
 
@@ -19,11 +23,12 @@ More information needed
19
 
20
  ## Intended uses & limitations
21
 
22
- More information needed
23
 
24
  ## Training and evaluation data
25
 
26
- More information needed
 
27
 
28
  ## Training procedure
29
 
@@ -43,4 +48,4 @@ The following hyperparameters were used during training:
43
  - Transformers 4.35.2
44
  - Pytorch 2.1.1+cu121
45
  - Datasets 2.15.0
46
- - Tokenizers 0.15.0
 
4
  model-index:
5
  - name: XLM_normalization_BEST_MODEL
6
  results: []
7
+ language:
8
+ - es
9
+ - en
10
+ - it
11
+ - fr
12
+ - eu
13
  ---
14
 
 
 
 
15
  # XLM_normalization_BEST_MODEL
16
 
17
+ This model was trained over the XLM-Large model for temporal expression normalization as a result of the paper "A Novel Methodology for Enhancing
18
+ Cross-Language and Domain Adaptability in Temporal Expression Normalization"
19
 
20
  ## Model description
21
 
 
23
 
24
  ## Intended uses & limitations
25
 
26
+ This model requires from extra post-processing. The proper code can be found at "https://github.com/asdc-s5/Temporal-expression-normalization-with-fill-mask"
27
 
28
  ## Training and evaluation data
29
 
30
+ All the information about training, evaluation and benchmarking can be found in the paper "A Novel Methodology for Enhancing
31
+ Cross-Language and Domain Adaptability in Temporal Expression Normalization"
32
 
33
  ## Training procedure
34
 
 
48
  - Transformers 4.35.2
49
  - Pytorch 2.1.1+cu121
50
  - Datasets 2.15.0
51
+ - Tokenizers 0.15.0