rulm
34 строки · 966.0 Байт
1{
2"trainer": {3"evaluation_strategy": "steps",4"per_device_train_batch_size": 4,5"per_device_eval_batch_size": 4,6"gradient_accumulation_steps": 32,7"eval_steps": 150,8"save_steps": 150,9"logging_steps": 5,10"learning_rate": 0.0003,11"num_train_epochs": 3,12"lr_scheduler_type": "cosine",13"warmup_steps": 100,14"fp16": true,15"bf16": false,16"torch_compile": false,17"optim": "adamw_torch"18},19"lora": {20"r": 16,21"lora_alpha": 16,22"lora_dropout": 0.05,23"bias": "none",24"target_modules": ["q_proj", "v_proj"],25"task_type": "CAUSAL_LM"26},27"load_in_8bit": true,28"only_target_loss": false,29"model_name": "models/llama-13b",30"templates_path": "internal_prompts/ru_alpaca.json",31"model_type": "causal",32"max_source_tokens_count": 512,33"max_target_tokens_count": 83234}
35