ardaorcun commited on
Commit
3e07768
·
verified ·
1 Parent(s): 964a601

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -30
README.md CHANGED
@@ -9,45 +9,49 @@ pipeline_tag: text-generation
9
 
10
  # Model Card for Model ID
11
 
12
- This model is finuted version of YTU's Cosmos GPT2 Language Model
13
 
14
  ## Training Details
15
 
16
- Model fine-tuned by using LoRA and QLoRA. Training parameters is defined below.
17
 
18
- LoRA configs:
19
 
20
- r=16,
21
- lora_alpha=32,
22
- target_modules=['c_proj',
23
- 'c_fc',
24
- 'gate_proj',
25
- 'c_proj',
26
- 'c_attn'],
27
- bias="lora_only",
28
- use_rslora=True,
29
- fan_in_fan_out=True,
30
- lora_dropout=0.05,
31
- task_type="CAUSAL_LM",
32
-
33
 
 
 
 
 
 
 
 
34
 
 
35
 
 
36
 
37
 
38
- Train Parameters:
39
- num_train_epochs=5,
40
- per_device_train_batch_size=10,
41
- gradient_accumulation_steps=1,
42
- gradient_checkpointing=True,
43
- optim="paged_lion_8bit",
44
- logging_steps=11,
45
- save_strategy="epoch",
46
- learning_rate=2e-4,
47
- max_grad_norm=0.3,
48
- warmup_ratio=0.03,
49
- lr_scheduler_type="linear"
 
 
50
 
51
- ### Training Data
52
 
53
- For training i used Merve's Turkish Instructions Dataset you can check here -> https://huggingface.co/datasets/merve/turkish_instructions
 
9
 
10
  # Model Card for Model ID
11
 
12
+ This model is a fine-tuned version of YTU's Cosmos GPT2 Language Model.
13
 
14
  ## Training Details
15
 
16
+ The model was fine-tuned using LoRA and QLoRA techniques. Training parameters are defined below.
17
 
18
+ ### LoRA configs:
19
 
20
+ - **r**=16
21
+ - **lora_alpha**=32
22
+ - **target_modules**=c_proj,c_fc, gate_proj, c_proj, c_attn
23
+ - **lora_dropout**=0.05
24
+ - **bias**="lora_only"
25
+ - **fan_in_fan_out**=True
26
+ - **max_seq_length**=512
27
+ - **use_rslora**=True
 
 
 
 
 
28
 
29
+ ### Train Parameters:
30
+ - **train_epochs**=5
31
+ - **optim**="paged_lion_8bit"
32
+ - **learning_rate**=2e-4
33
+ - **warmup_ratio**=0.03
34
+ - **max_grad_norm**=0.3
35
+ - **lr_scheduler_type**="linear"
36
 
37
+ ### Training Data
38
 
39
+ For training, I used Merve's Turkish Instructions Dataset, which you can check here: <a href="https://huggingface.co/datasets/merve/turkish_instructions">Merve's Turkish Instructions Dataset</a>
40
 
41
 
42
+ ## Instruction template:
43
+
44
+ ```python
45
+ def format_instruction(sample):
46
+ return f"""Sen cevap vermeyi seven yardımcı bir dil modelisin.
47
+ ### Input:
48
+ {sample["talimat"]}
49
+
50
+ ### Context:
51
+ {sample[" giriş"]}
52
+
53
+ ### Response:
54
+ {sample[" çıktı"]}
55
+ """```
56
 
 
57