sgraham commited on
Commit
2e27bbb
·
1 Parent(s): a0611cd

trained on 1371 images from the Met for 10 epochs

Browse files

***** train metrics *****
epoch = 10.0
train_loss = 0.8552
train_runtime = 1:11:30.94
train_samples_per_second = 3.225
train_steps_per_second = 0.033

Files changed (2) hide show
  1. config.json +22 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/clip-vit-base-patch32",
3
+ "architectures": [
4
+ "CLIPModel"
5
+ ],
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 512,
10
+ "text_config": {
11
+ "bos_token_id": 0,
12
+ "dropout": 0.0,
13
+ "eos_token_id": 2,
14
+ "model_type": "clip_text_model"
15
+ },
16
+ "torch_dtype": "float32",
17
+ "transformers_version": "4.35.0.dev0",
18
+ "vision_config": {
19
+ "dropout": 0.0,
20
+ "model_type": "clip_vision_model"
21
+ }
22
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b21f784f4b57697ba0c977c2bc9e890ab8c992157fcbc4f9cc573484783fc03
3
+ size 605240770