You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

config.yaml 2.1KB

3 months ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105
  1. shared:
  2. project_name: continual_prompt_pretrained_mlp
  3. use_tqdm: true
  4. random_seed: 42
  5. default: &default
  6. model_name: google/t5-large-lm-adapt
  7. wandb_name: null
  8. train_batch_size: 32
  9. valid_batch_size: 32
  10. num_epochs: 100
  11. peft_params: null # no mutation
  12. hot_modules: null # fine-tune all
  13. balancify_train: false
  14. best_finder:
  15. save: true
  16. metric: valid_f1-score-ma
  17. higher_better: true
  18. tasks:
  19. - glue:cola
  20. run_configs:
  21. # - <<: *default
  22. # wandb_name: large_5t_mlp128
  23. # learning_rate: 0.02
  24. # hot_modules:
  25. # - sadcl_learned_embeddin
  26. # train_batch_size: 24
  27. # valid_batch_size: 24
  28. # peft_params:
  29. # kind: encoder_emb
  30. # n_tokens: 5
  31. # mlp_emb: 128
  32. # - <<: *default
  33. # wandb_name: large_10t_mlp128
  34. # learning_rate: 0.02
  35. # hot_modules:
  36. # - sadcl_learned_embeddin
  37. # train_batch_size: 24
  38. # valid_batch_size: 24
  39. # peft_params:
  40. # kind: encoder_emb
  41. # n_tokens: 10
  42. # mlp_emb: 128
  43. # - <<: *default
  44. # wandb_name: large_5t_mlp128_not_freeze
  45. # learning_rate: 0.02
  46. # hot_modules:
  47. # - sadcl
  48. # train_batch_size: 24
  49. # valid_batch_size: 24
  50. # peft_params:
  51. # kind: encoder_emb
  52. # n_tokens: 5
  53. # mlp_emb: 128
  54. # - <<: *default
  55. # wandb_name: large_10t_mlp128_not_freeze
  56. # learning_rate: 0.02
  57. # hot_modules:
  58. # - sadcl
  59. # train_batch_size: 24
  60. # valid_batch_size: 24
  61. # peft_params:
  62. # kind: encoder_emb
  63. # n_tokens: 10
  64. # mlp_emb: 128
  65. # - <<: *default
  66. # wandb_name: large_5t_mlp128_not_freeze_lowlr
  67. # learning_rate: 0.001
  68. # hot_modules:
  69. # - sadcl
  70. # train_batch_size: 24
  71. # valid_batch_size: 24
  72. # peft_params:
  73. # kind: encoder_emb
  74. # n_tokens: 5
  75. # mlp_emb: 128
  76. # - <<: *default
  77. # wandb_name: large_10t_mlp128_not_freeze_lowlr
  78. # learning_rate: 0.001
  79. # hot_modules:
  80. # - sadcl
  81. # train_batch_size: 24
  82. # valid_batch_size: 24
  83. # peft_params:
  84. # kind: encoder_emb
  85. # n_tokens: 10
  86. # mlp_emb: 128
  87. - <<: *default
  88. wandb_name: large_100t_mlp128_lr.02
  89. learning_rate: 0.02
  90. hot_modules:
  91. - sadcl_learned_embeddin
  92. train_batch_size: 24
  93. valid_batch_size: 24
  94. peft_params:
  95. kind: encoder_emb
  96. n_tokens: 100
  97. mlp_emb: 128