|
1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162 |
- shared:
- project_name: lowdim_prompts
- use_tqdm: true
- random_seed: 42
-
- default: &default
- model_name: google/t5-large-lm-adapt
- wandb_name: null
- train_batch_size: 32
- valid_batch_size: 32
- num_epochs: 200
- peft_params: null # no mutation
- hot_modules: null # fine-tune all
- balancify_train: false
- best_finder:
- save: true
- metric: valid_f1-score-ma
- higher_better: true
- tasks:
- - glue:cola
-
- run_configs:
- # - <<: *default
- # wandb_name: n_tokens100_n_comb_tokens512
- # learning_rate: 0.01
- # hot_modules:
- # - sadcl
- # peft_params:
- # kind: comb_prompt
- # n_tokens: 100
- # n_comb_tokens: 512
- # - <<: *default
- # wandb_name: n_tokens100_n_comb_tokens2048
- # learning_rate: 0.01
- # hot_modules:
- # - sadcl
- # peft_params:
- # kind: comb_prompt
- # n_tokens: 100
- # n_comb_tokens: 2048
- - <<: *default
- wandb_name: large_n_tokens100_64_256
- learning_rate: 0.01
- hot_modules:
- - sadcl
- peft_params:
- kind: lowdim_prompt
- n_tokens: 100
- dims:
- - 64
- - 256
- - <<: *default
- wandb_name: large_n_tokens100_256_512
- learning_rate: 0.01
- hot_modules:
- - sadcl
- peft_params:
- kind: lowdim_prompt
- n_tokens: 100
- dims:
- - 256
- - 512
|