cjmijones commited on
Commit
fb7fe98
·
verified ·
1 Parent(s): 6c38174

Training in progress, epoch 3

Browse files
README.md CHANGED
@@ -1,15 +1,15 @@
1
  ---
2
  base_model: google/gemma-3-4b-pt
3
  library_name: transformers
4
- model_name: gemma_text_to_sql_run_20251208_051516
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
  - trl
 
9
  licence: license
10
  ---
11
 
12
- # Model Card for gemma_text_to_sql_run_20251208_051516
13
 
14
  This model is a fine-tuned version of [google/gemma-3-4b-pt](https://huggingface.co/google/gemma-3-4b-pt).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
1
  ---
2
  base_model: google/gemma-3-4b-pt
3
  library_name: transformers
4
+ model_name: gemma_text_to_sql_run_20251208_052804
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
12
+ # Model Card for gemma_text_to_sql_run_20251208_052804
13
 
14
  This model is a fine-tuned version of [google/gemma-3-4b-pt](https://huggingface.co/google/gemma-3-4b-pt).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
adapter_config.json CHANGED
@@ -12,7 +12,7 @@
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
- "lora_alpha": 32,
16
  "lora_bias": false,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
@@ -22,17 +22,17 @@
22
  "embed_tokens"
23
  ],
24
  "peft_type": "LORA",
25
- "r": 64,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
- "o_proj",
30
  "down_proj",
 
31
  "gate_proj",
32
- "up_proj",
33
- "v_proj",
34
  "k_proj",
35
- "q_proj"
 
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
 
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
+ "lora_alpha": 16,
16
  "lora_bias": false,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
 
22
  "embed_tokens"
23
  ],
24
  "peft_type": "LORA",
25
+ "r": 32,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "v_proj",
30
  "down_proj",
31
+ "o_proj",
32
  "gate_proj",
 
 
33
  "k_proj",
34
+ "q_proj",
35
+ "up_proj"
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97ff7132eb161e6dad4b2ba135768fd4767859a68b7e090703926c12bf414cf0
3
- size 3209724888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:156e9816eae1f0195408e0059ea3ddcebbba24caa0df3f3f199aa35d248039d0
3
+ size 2947416824
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f0ee528da3de243812c35efee45b1f4e6c10d4678d887b4d052b3b6285d1e53
3
  size 6289
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53cabff3e69bb6d1432492443d3386bff6832acca5209780eeaf5b4dd386e9f3
3
  size 6289