LysandreJik commited on
Commit
4967704
Β·
1 Parent(s): 0057c8b

Take model from serialization

Browse files
.gitattributes CHANGED
@@ -16,3 +16,5 @@
16
  *.pth filter=lfs diff=lfs merge=lfs -text
17
  serialization/model.tar.gz filter=lfs diff=lfs merge=lfs -text
18
  serialization/best.th filter=lfs diff=lfs merge=lfs -text
 
 
 
16
  *.pth filter=lfs diff=lfs merge=lfs -text
17
  serialization/model.tar.gz filter=lfs diff=lfs merge=lfs -text
18
  serialization/best.th filter=lfs diff=lfs merge=lfs -text
19
+ weights.th filter=lfs diff=lfs merge=lfs -text
20
+ model.tar.gz filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_reader": {
3
+ "type": "conll2003",
4
+ "tag_label": "ner",
5
+ "token_indexers": {
6
+ "tokens": {
7
+ "type": "single_id",
8
+ "lowercase_tokens": true
9
+ }
10
+ }
11
+ },
12
+ "model": {
13
+ "type": "simple_tagger",
14
+ "calculate_span_f1": true,
15
+ "encoder": {
16
+ "type": "lstm",
17
+ "hidden_size": 4,
18
+ "input_size": 2,
19
+ "num_layers": 1
20
+ },
21
+ "label_encoding": "IOB1",
22
+ "text_field_embedder": {
23
+ "token_embedders": {
24
+ "tokens": {
25
+ "type": "embedding",
26
+ "embedding_dim": 100,
27
+ "pretrained_file": "allennlp/tests/fixtures/embeddings/glove.6B.100d.sample.txt.gz",
28
+ "projection_dim": 2,
29
+ "trainable": true
30
+ }
31
+ }
32
+ }
33
+ },
34
+ "train_data_path": "allennlp/tests/fixtures/data/conll2003.txt",
35
+ "validation_data_path": "allennlp/tests/fixtures/data/conll2003.txt",
36
+ "trainer": {
37
+ "cuda_device": -1,
38
+ "grad_norm": 1,
39
+ "num_epochs": 1,
40
+ "optimizer": {
41
+ "type": "adadelta",
42
+ "lr": 1e-06,
43
+ "rho": 0.95
44
+ },
45
+ "patience": 500
46
+ },
47
+ "data_loader": {
48
+ "batch_sampler": {
49
+ "type": "bucket",
50
+ "batch_size": 80,
51
+ "padding_noise": 0
52
+ }
53
+ },
54
+ "validation_dataset_reader": {
55
+ "type": "conll2003",
56
+ "tag_label": "ner",
57
+ "token_indexers": {
58
+ "tokens": {
59
+ "type": "single_id",
60
+ "lowercase_tokens": true,
61
+ "namespace": "test_tokens"
62
+ }
63
+ }
64
+ }
65
+ }
experiment.json DELETED
@@ -1,65 +0,0 @@
1
- {
2
- "dataset_reader": {
3
- "type": "conll2003",
4
- "tag_label": "ner",
5
- "token_indexers": {
6
- "tokens": {
7
- "type": "single_id",
8
- "lowercase_tokens": true
9
- }
10
- }
11
- },
12
- "validation_dataset_reader": {
13
- "type": "conll2003",
14
- "tag_label": "ner",
15
- "token_indexers": {
16
- "tokens": {
17
- "type": "single_id",
18
- "namespace": "test_tokens",
19
- "lowercase_tokens": true
20
- }
21
- }
22
- },
23
- "train_data_path": "test_fixtures/data/conll2003.txt",
24
- "validation_data_path": "test_fixtures/data/conll2003.txt",
25
- "model": {
26
- "type": "simple_tagger",
27
- "calculate_span_f1": true,
28
- "label_encoding": "IOB1",
29
- "text_field_embedder": {
30
- "token_embedders": {
31
- "tokens": {
32
- "type": "embedding",
33
- "projection_dim": 2,
34
- "pretrained_file": "test_fixtures/embeddings/glove.6B.100d.sample.txt.gz",
35
- "embedding_dim": 100,
36
- "trainable": true
37
- }
38
- }
39
- },
40
- "encoder": {
41
- "type": "lstm",
42
- "input_size": 2,
43
- "hidden_size": 4,
44
- "num_layers": 1
45
- }
46
- },
47
- "data_loader": {
48
- "batch_sampler":{
49
- "type": "bucket",
50
- "padding_noise": 0.0,
51
- "batch_size" : 80
52
- }
53
- },
54
- "trainer": {
55
- "num_epochs": 1,
56
- "grad_norm": 1.0,
57
- "patience": 500,
58
- "cuda_device": -1,
59
- "optimizer": {
60
- "type": "adadelta",
61
- "lr": 0.000001,
62
- "rho": 0.95
63
- }
64
- }
65
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
serialization/model.tar.gz DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c281875582a0dd7c9bdf2af74f0316536fc38651bc5cabf3395c6df67d737085
3
- size 9015
 
 
 
 
{serialization/vocabulary β†’ vocabulary}/labels.txt RENAMED
File without changes
{serialization/vocabulary β†’ vocabulary}/non_padded_namespaces.txt RENAMED
File without changes
{serialization/vocabulary β†’ vocabulary}/test_tokens.txt RENAMED
File without changes
{serialization/vocabulary β†’ vocabulary}/tokens.txt RENAMED
File without changes
serialization/best.th β†’ weights.th RENAMED
File without changes