Meta Byte Track
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

meta_trainer.py 13KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387
  1. # Mahdi Abdollahpour, 27/11/2021, 02:35 PM, PyCharm, ByteTrack
  2. from loguru import logger
  3. import torch
  4. from torch.nn.parallel import DistributedDataParallel as DDP
  5. from torch.utils.tensorboard import SummaryWriter
  6. from yolox.data import DataPrefetcher
  7. from yolox.utils import (
  8. MeterBuffer,
  9. ModelEMA,
  10. all_reduce_norm,
  11. get_model_info,
  12. get_rank,
  13. get_world_size,
  14. gpu_mem_usage,
  15. load_ckpt,
  16. occupy_mem,
  17. save_checkpoint,
  18. setup_logger,
  19. synchronize
  20. )
  21. import datetime
  22. import os
  23. import time
  24. import learn2learn as l2l
  25. class MetaTrainer:
  26. def __init__(self, exp, args):
  27. # init function only defines some basic attr, other attrs like model, optimizer are built in
  28. # before_train methods.
  29. self.exp = exp
  30. self.args = args
  31. # training related attr
  32. self.max_epoch = exp.max_epoch
  33. self.amp_training = args.fp16
  34. self.scaler = torch.cuda.amp.GradScaler(enabled=args.fp16)
  35. self.is_distributed = get_world_size() > 1
  36. self.rank = get_rank()
  37. self.local_rank = args.local_rank
  38. self.device = "cuda:{}".format(self.local_rank)
  39. self.use_model_ema = exp.ema
  40. # data/dataloader related attr
  41. self.data_type = torch.float16 if args.fp16 else torch.float32
  42. self.input_size = exp.input_size
  43. self.best_ap = 0
  44. self.adaptation_period = args.adaptation_period
  45. # metric record
  46. self.meter = MeterBuffer(window_size=exp.print_interval)
  47. self.file_name = os.path.join(exp.output_dir, args.experiment_name)
  48. if self.rank == 0:
  49. os.makedirs(self.file_name, exist_ok=True)
  50. setup_logger(
  51. self.file_name,
  52. distributed_rank=self.rank,
  53. filename="train_log.txt",
  54. mode="a",
  55. )
  56. def train(self):
  57. self.before_train()
  58. try:
  59. self.train_in_epoch()
  60. except Exception:
  61. raise
  62. finally:
  63. self.after_train()
  64. def train_in_epoch(self):
  65. for self.epoch in range(self.start_epoch, self.max_epoch):
  66. self.before_epoch()
  67. self.train_in_task()
  68. self.after_epoch()
  69. def train_in_iter(self):
  70. for self.iter in range(len(self.train_loader)):
  71. self.before_iter()
  72. self.train_one_iter()
  73. self.after_iter()
  74. def train_in_task(self):
  75. for i in range(len(self.train_loaders)):
  76. self.before_task(i)
  77. self.train_in_iter()
  78. self.after_task(i)
  79. def before_task(self, i):
  80. # logger.info("init prefetcher, this might take one minute or less...")
  81. self.train_loader = self.train_loaders[i]
  82. self.prefetcher = self.prefetchers[i]
  83. self.learner = self.model.clone()
  84. logger.info("model clone created. dataloader:{}".format(i))
  85. def after_task(self,i):
  86. self.epoch_progress += len(self.train_loaders[i])
  87. def adapt(self, inps, targets):
  88. # adapt_inps =inps[:1, ...]
  89. # targets_inps =targets[:1, ...]
  90. # print(adapt_inps.shape)
  91. # print(targets_inps.shape)
  92. outputs = self.learner(inps, targets)
  93. loss = outputs["total_loss"]
  94. self.learner.adapt(loss)
  95. del outputs, loss
  96. def train_one_iter(self):
  97. iter_start_time = time.time()
  98. inps, targets = self.prefetcher.next()
  99. inps = inps.to(self.data_type)
  100. targets = targets.to(self.data_type)
  101. targets.requires_grad = False
  102. data_end_time = time.time()
  103. # logger.info("in train iter")
  104. with torch.cuda.amp.autocast(enabled=self.amp_training):
  105. if self.iter % self.adaptation_period == 0:
  106. self.adapt(inps, targets)
  107. outputs = self.learner(inps, targets)
  108. loss = outputs["total_loss"]
  109. for p in self.exp.all_parameters:
  110. if p.grad is not None:
  111. p.grad.data.mul_(1.0 / self.args.batch_size)
  112. self.optimizer.zero_grad()
  113. self.scaler.scale(loss).backward()
  114. self.scaler.step(self.optimizer)
  115. self.scaler.update()
  116. if self.use_model_ema:
  117. self.ema_model.update(self.model)
  118. lr = self.lr_scheduler.update_lr(self.progress_in_iter + 1)
  119. for param_group in self.optimizer.param_groups:
  120. param_group["lr"] = lr
  121. iter_end_time = time.time()
  122. self.meter.update(
  123. iter_time=iter_end_time - iter_start_time,
  124. data_time=data_end_time - iter_start_time,
  125. lr=lr,
  126. **outputs,
  127. )
  128. del loss, outputs
  129. def before_train(self):
  130. logger.info("args: {}".format(self.args))
  131. # logger.info("exp value:\n{}".format(self.exp))
  132. # model related init
  133. torch.cuda.set_device(self.local_rank)
  134. model = self.exp.get_model()
  135. logger.info(
  136. "Model Summary: {}".format(get_model_info(model, self.exp.test_size))
  137. )
  138. # exit()
  139. model.to(self.device)
  140. # from torchsummary import summary
  141. # summary(model, input_size=(3, 300, 300), device='cuda')
  142. # value of epoch will be set in `resume_train`
  143. self.model = l2l.algorithms.MAML(model, lr=self.exp.inner_lr, first_order=self.exp.first_order)
  144. # solver related init
  145. self.optimizer = self.exp.get_optimizer(self.args.batch_size)
  146. self.model = self.resume_train(self.model)
  147. # data related init
  148. self.no_aug = self.start_epoch >= self.max_epoch - self.exp.no_aug_epochs
  149. logger.info('Getting data loaders')
  150. self.train_loaders = self.exp.get_data_loaders(
  151. batch_size=self.args.batch_size,
  152. is_distributed=self.is_distributed,
  153. no_aug=self.no_aug,
  154. )
  155. # max_iter means iters per epoch
  156. self.max_iter = 0
  157. self.prefetchers = []
  158. for i, train_loader in enumerate(self.train_loaders):
  159. self.max_iter += len(train_loader)
  160. self.prefetchers.append(DataPrefetcher(train_loader))
  161. self.lr_scheduler = self.exp.get_lr_scheduler(
  162. self.exp.basic_lr_per_img * self.args.batch_size, self.max_iter
  163. )
  164. if self.args.occupy:
  165. occupy_mem(self.local_rank)
  166. if self.is_distributed:
  167. self.model = DDP(self.model, device_ids=[self.local_rank], broadcast_buffers=False)
  168. if self.use_model_ema:
  169. self.ema_model = ModelEMA(self.model, 0.9998)
  170. self.ema_model.updates = self.max_iter * self.start_epoch
  171. # self.model = model
  172. self.model.train()
  173. self.evaluator = self.exp.get_evaluator(
  174. batch_size=self.args.batch_size, is_distributed=self.is_distributed
  175. )
  176. # Tensorboard logger
  177. if self.rank == 0:
  178. self.tblogger = SummaryWriter(self.file_name)
  179. logger.info("Training start...")
  180. # logger.info("\n{}".format(model))
  181. def after_train(self):
  182. logger.info(
  183. "Training of experiment is done and the best AP is {:.2f}".format(
  184. self.best_ap * 100
  185. )
  186. )
  187. def before_epoch(self):
  188. logger.info("---> start train epoch{}".format(self.epoch + 1))
  189. self.epoch_progress = 0
  190. if self.epoch + 1 == self.max_epoch - self.exp.no_aug_epochs or self.no_aug:
  191. logger.info("--->No mosaic aug now!")
  192. for train_loader in self.train_loaders:
  193. train_loader.close_mosaic()
  194. logger.info("--->Add additional L1 loss now!")
  195. if self.is_distributed:
  196. self.model.module.head.use_l1 = True
  197. else:
  198. self.model.head.use_l1 = True
  199. self.exp.eval_interval = 1
  200. if not self.no_aug:
  201. self.save_ckpt(ckpt_name="last_mosaic_epoch")
  202. def after_epoch(self):
  203. self.epoch_progress = 0
  204. if self.use_model_ema:
  205. self.ema_model.update_attr(self.model)
  206. self.save_ckpt(ckpt_name="latest")
  207. if (self.epoch + 1) % self.exp.eval_interval == 0:
  208. all_reduce_norm(self.model)
  209. self.evaluate_and_save_model()
  210. def before_iter(self):
  211. pass
  212. def after_iter(self):
  213. """
  214. `after_iter` contains two parts of logic:
  215. * log information
  216. * reset setting of resize
  217. """
  218. # log needed information
  219. if (self.iter + 1) % self.exp.print_interval == 0:
  220. # TODO check ETA logic
  221. left_iters = self.max_iter * self.max_epoch - (self.progress_in_iter + 1)
  222. eta_seconds = self.meter["iter_time"].global_avg * left_iters
  223. eta_str = "ETA: {}".format(datetime.timedelta(seconds=int(eta_seconds)))
  224. progress_str = "epoch: {}/{}, iter: {}/{}".format(
  225. self.epoch + 1, self.max_epoch, self.iter + 1 + self.epoch_progress, self.max_iter
  226. )
  227. loss_meter = self.meter.get_filtered_meter("loss")
  228. loss_str = ", ".join(
  229. ["{}: {:.3f}".format(k, v.latest) for k, v in loss_meter.items()]
  230. )
  231. time_meter = self.meter.get_filtered_meter("time")
  232. time_str = ", ".join(
  233. ["{}: {:.3f}s".format(k, v.avg) for k, v in time_meter.items()]
  234. )
  235. logger.info(
  236. "{}, mem: {:.0f}Mb, {}, {}, lr: {:.3e}".format(
  237. progress_str,
  238. gpu_mem_usage(),
  239. time_str,
  240. loss_str,
  241. self.meter["lr"].latest,
  242. )
  243. + (", size: {:d}, {}".format(self.input_size[0], eta_str))
  244. )
  245. self.meter.clear_meters()
  246. # random resizing
  247. if self.exp.random_size is not None and (self.progress_in_iter + 1) % 10 == 0:
  248. self.input_size = self.exp.random_resize(
  249. self.train_loader, self.epoch, self.rank, self.is_distributed
  250. )
  251. @property
  252. def progress_in_iter(self):
  253. return self.epoch * self.max_iter + self.iter + self.epoch_progress
  254. def resume_train(self, model):
  255. if self.args.resume:
  256. logger.info("resume training")
  257. if self.args.ckpt is None:
  258. ckpt_file = os.path.join(self.file_name, "latest" + "_ckpt.pth.tar")
  259. else:
  260. ckpt_file = self.args.ckpt
  261. ckpt = torch.load(ckpt_file, map_location=self.device)
  262. # handling meta models
  263. # new_dict = {}
  264. # for key in ckpt["model"].keys():
  265. # if key.startswith('module.'):
  266. # new_dict[key[7:]] = ckpt["model"][key]
  267. # else:
  268. # new_dict[key] = ckpt["model"][key]
  269. # del ckpt["model"]
  270. # ckpt["model"] = new_dict
  271. # resume the model/optimizer state dict
  272. model.load_state_dict(ckpt["model"])
  273. self.optimizer.load_state_dict(ckpt["optimizer"])
  274. start_epoch = (
  275. self.args.start_epoch - 1
  276. if self.args.start_epoch is not None
  277. else ckpt["start_epoch"]
  278. )
  279. self.start_epoch = start_epoch
  280. logger.info(
  281. "loaded checkpoint '{}' (epoch {})".format(
  282. self.args.resume, self.start_epoch
  283. )
  284. ) # noqa
  285. else:
  286. if self.args.ckpt is not None:
  287. logger.info("loading checkpoint for fine tuning")
  288. ckpt_file = self.args.ckpt
  289. ckpt = torch.load(ckpt_file, map_location=self.device)["model"]
  290. model = load_ckpt(model, ckpt)
  291. self.start_epoch = 0
  292. return model
  293. def evaluate_and_save_model(self):
  294. evalmodel = self.ema_model.ema if self.use_model_ema else self.model
  295. ap50_95, ap50, summary = self.exp.eval(
  296. evalmodel, self.evaluator, self.is_distributed
  297. )
  298. self.model.train()
  299. if self.rank == 0:
  300. self.tblogger.add_scalar("val/COCOAP50", ap50, self.epoch + 1)
  301. self.tblogger.add_scalar("val/COCOAP50_95", ap50_95, self.epoch + 1)
  302. logger.info("\n" + summary)
  303. synchronize()
  304. # self.best_ap = max(self.best_ap, ap50_95)
  305. self.save_ckpt("last_epoch", ap50 > self.best_ap)
  306. self.best_ap = max(self.best_ap, ap50)
  307. def save_ckpt(self, ckpt_name, update_best_ckpt=False):
  308. if self.rank == 0:
  309. save_model = self.ema_model.ema if self.use_model_ema else self.model
  310. logger.info("Save weights to {}".format(self.file_name))
  311. ckpt_state = {
  312. "start_epoch": self.epoch + 1,
  313. "model": save_model.state_dict(),
  314. "optimizer": self.optimizer.state_dict(),
  315. }
  316. save_checkpoint(
  317. ckpt_state,
  318. update_best_ckpt,
  319. self.file_name,
  320. ckpt_name,
  321. )