huseinzol05 commited on
Commit
c6bba1f
1 Parent(s): 5a527dc

Upload MM_LLMs

Browse files
Files changed (2) hide show
  1. config.json +3 -9
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "_name_or_path": "multimodal-tinyllama-whisper-small-siglip-v1/checkpoint-16700",
3
  "architectures": [
4
  "MM_LLMs"
5
  ],
6
- "attention_heads": 8,
7
  "audio_config": {
8
  "_name_or_path": "mesolitica/malaysian-whisper-small",
9
  "activation_dropout": 0.0,
@@ -204,11 +203,9 @@
204
  "use_weighted_layer_sum": false,
205
  "vocab_size": 51865
206
  },
207
- "audio_conv_kernel": 240,
208
- "audio_conv_stride": 220,
209
  "hidden_size": 2048,
210
  "image_config": {
211
- "_name_or_path": "google/siglip-base-patch16-224",
212
  "add_cross_attention": false,
213
  "architectures": [
214
  "SiglipModel"
@@ -370,7 +367,7 @@
370
  "0": "LABEL_0",
371
  "1": "LABEL_1"
372
  },
373
- "image_size": 224,
374
  "intermediate_size": 3072,
375
  "is_decoder": false,
376
  "is_encoder_decoder": false,
@@ -418,8 +415,6 @@
418
  "use_bfloat16": false
419
  }
420
  },
421
- "image_conv_kernel": 48,
422
- "image_conv_stride": 36,
423
  "llm_config": {
424
  "_name_or_path": "mesolitica/malaysian-tinyllama-1.1b-16k-instructions-v3",
425
  "add_cross_attention": false,
@@ -502,7 +497,6 @@
502
  "vocab_size": 32004
503
  },
504
  "model_type": "mm_llms",
505
- "n_frames": 6,
506
  "torch_dtype": "bfloat16",
507
  "transformers_version": "4.37.1"
508
  }
 
1
  {
2
+ "_name_or_path": "multimodal-tinyllama-whisper-small-siglip/checkpoint-400",
3
  "architectures": [
4
  "MM_LLMs"
5
  ],
 
6
  "audio_config": {
7
  "_name_or_path": "mesolitica/malaysian-whisper-small",
8
  "activation_dropout": 0.0,
 
203
  "use_weighted_layer_sum": false,
204
  "vocab_size": 51865
205
  },
 
 
206
  "hidden_size": 2048,
207
  "image_config": {
208
+ "_name_or_path": "google/siglip-base-patch16-384",
209
  "add_cross_attention": false,
210
  "architectures": [
211
  "SiglipModel"
 
367
  "0": "LABEL_0",
368
  "1": "LABEL_1"
369
  },
370
+ "image_size": 384,
371
  "intermediate_size": 3072,
372
  "is_decoder": false,
373
  "is_encoder_decoder": false,
 
415
  "use_bfloat16": false
416
  }
417
  },
 
 
418
  "llm_config": {
419
  "_name_or_path": "mesolitica/malaysian-tinyllama-1.1b-16k-instructions-v3",
420
  "add_cross_attention": false,
 
497
  "vocab_size": 32004
498
  },
499
  "model_type": "mm_llms",
 
500
  "torch_dtype": "bfloat16",
501
  "transformers_version": "4.37.1"
502
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c00b609cae60891f2a3c1dbbd123dc28ad4cc0f1f799f7fcb7da6666e7e7a58
3
- size 3504434068
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4805b652684d9bb425f952ec1e95324e7ddafa2ad5e3d7266fca43c3652ec60
3
+ size 3236402604