Simran Arora commited on
Commit
77c059e
1 Parent(s): bba5f57
Files changed (2) hide show
  1. config.json +1 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"d_model": 2048, "n_layer": 46, "vocab_size": 50277, "ssm_cfg": {}, "rms_norm": true, "residual_in_fp32": true, "fused_add_norm": true, "pad_vocab_size_multiple": 16, "reorder_and_upcast_attn": false, "scale_attn_by_inverse_layer_idx": true, "n_positions": 2048, "n_embd": 1024, "n_head": 16, "use_flash_attn": true, "fused_dropout_add_ln": true, "fused_mlp": true, "fused_bias_fc": true, "use_fast_path": true}
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4eda5f28e8be0e17e64712619bab7ee6bcbce5bb6f07ebc7f52315ef3dcee4c
3
+ size 5277401485