nklockiewicz commited on
Commit
ed48091
·
verified ·
1 Parent(s): 3ff3d74

Create quantized/config.json

Browse files
Files changed (1) hide show
  1. quantized/config.json +97 -0
quantized/config.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Lfm2VlForConditionalGeneration"
4
+ ],
5
+ "do_image_splitting": true,
6
+ "downsample_factor": 2,
7
+ "dtype": "bfloat16",
8
+ "encoder_patch_size": 16,
9
+ "image_token_id": 396,
10
+ "max_image_tokens": 256,
11
+ "max_pixels_tolerance": 2.0,
12
+ "max_tiles": 10,
13
+ "min_image_tokens": 64,
14
+ "min_tiles": 2,
15
+ "model_type": "lfm2_vl",
16
+ "projector_bias": true,
17
+ "projector_hidden_act": "gelu",
18
+ "projector_hidden_size": 2048,
19
+ "projector_use_layernorm": false,
20
+ "text_config": {
21
+ "_name_or_path": "LiquidAI/LFM2-1.2B",
22
+ "architectures": [
23
+ "Lfm2ForCausalLM"
24
+ ],
25
+ "block_auto_adjust_ff_dim": true,
26
+ "block_dim": 2048,
27
+ "block_ff_dim": 12288,
28
+ "block_ffn_dim_multiplier": 1.0,
29
+ "block_mlp_init_scale": 1.0,
30
+ "block_multiple_of": 256,
31
+ "block_norm_eps": 1e-05,
32
+ "block_out_init_scale": 1.0,
33
+ "block_use_swiglu": true,
34
+ "block_use_xavier_init": true,
35
+ "conv_L_cache": 3,
36
+ "conv_bias": false,
37
+ "conv_dim": 2048,
38
+ "conv_dim_out": 2048,
39
+ "conv_use_xavier_init": true,
40
+ "dtype": "bfloat16",
41
+ "eos_token_id": 7,
42
+ "hidden_size": 2048,
43
+ "initializer_range": 0.02,
44
+ "intermediate_size": 12288,
45
+ "layer_types": [
46
+ "conv",
47
+ "conv",
48
+ "full_attention",
49
+ "conv",
50
+ "conv",
51
+ "full_attention",
52
+ "conv",
53
+ "conv",
54
+ "full_attention",
55
+ "conv",
56
+ "full_attention",
57
+ "conv",
58
+ "full_attention",
59
+ "conv",
60
+ "full_attention",
61
+ "conv"
62
+ ],
63
+ "max_position_embeddings": 128000,
64
+ "model_type": "lfm2",
65
+ "norm_eps": 1e-05,
66
+ "num_attention_heads": 32,
67
+ "num_heads": 32,
68
+ "num_hidden_layers": 16,
69
+ "num_key_value_heads": 8,
70
+ "rope_parameters": {
71
+ "rope_theta": 1000000.0,
72
+ "rope_type": "default"
73
+ },
74
+ "use_cache": true,
75
+ "use_pos_enc": true,
76
+ "vocab_size": 65536
77
+ },
78
+ "tile_size": 512,
79
+ "transformers_version": "5.0.0.dev0",
80
+ "use_image_special_tokens": true,
81
+ "use_thumbnail": true,
82
+ "vision_config": {
83
+ "attention_dropout": 0.0,
84
+ "dtype": "bfloat16",
85
+ "hidden_act": "gelu_pytorch_tanh",
86
+ "hidden_size": 1152,
87
+ "intermediate_size": 4304,
88
+ "layer_norm_eps": 1e-06,
89
+ "model_type": "siglip2_vision_model",
90
+ "num_attention_heads": 16,
91
+ "num_channels": 3,
92
+ "num_hidden_layers": 27,
93
+ "num_patches": 256,
94
+ "patch_size": 16,
95
+ "vision_use_head": false
96
+ }
97
+ }