Meta Byte Track
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

meta_trainer.py 12KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368
  1. # Mahdi Abdollahpour, 27/11/2021, 02:35 PM, PyCharm, ByteTrack
  2. from loguru import logger
  3. import torch
  4. from torch.nn.parallel import DistributedDataParallel as DDP
  5. from torch.utils.tensorboard import SummaryWriter
  6. from yolox.data import DataPrefetcher
  7. from yolox.utils import (
  8. MeterBuffer,
  9. ModelEMA,
  10. all_reduce_norm,
  11. get_model_info,
  12. get_rank,
  13. get_world_size,
  14. gpu_mem_usage,
  15. load_ckpt,
  16. occupy_mem,
  17. save_checkpoint,
  18. setup_logger,
  19. synchronize
  20. )
  21. import datetime
  22. import os
  23. import time
  24. import learn2learn as l2l
  25. class MetaTrainer:
  26. def __init__(self, exp, args):
  27. # init function only defines some basic attr, other attrs like model, optimizer are built in
  28. # before_train methods.
  29. self.exp = exp
  30. self.args = args
  31. # training related attr
  32. self.max_epoch = exp.max_epoch
  33. self.amp_training = args.fp16
  34. self.scaler = torch.cuda.amp.GradScaler(enabled=args.fp16)
  35. self.is_distributed = get_world_size() > 1
  36. self.rank = get_rank()
  37. self.local_rank = args.local_rank
  38. self.device = "cuda:{}".format(self.local_rank)
  39. self.use_model_ema = exp.ema
  40. # data/dataloader related attr
  41. self.data_type = torch.float16 if args.fp16 else torch.float32
  42. self.input_size = exp.input_size
  43. self.best_ap = 0
  44. self.adaptation_period = args.adaptation_period
  45. # metric record
  46. self.meter = MeterBuffer(window_size=exp.print_interval)
  47. self.file_name = os.path.join(exp.output_dir, args.experiment_name)
  48. if self.rank == 0:
  49. os.makedirs(self.file_name, exist_ok=True)
  50. setup_logger(
  51. self.file_name,
  52. distributed_rank=self.rank,
  53. filename="train_log.txt",
  54. mode="a",
  55. )
  56. def train(self):
  57. self.before_train()
  58. try:
  59. self.train_in_epoch()
  60. except Exception:
  61. raise
  62. finally:
  63. self.after_train()
  64. def train_in_epoch(self):
  65. for self.epoch in range(self.start_epoch, self.max_epoch):
  66. self.before_epoch()
  67. self.train_in_task()
  68. self.after_epoch()
  69. def train_in_iter(self, task):
  70. for self.iter in range(len(task)):
  71. self.before_iter()
  72. self.train_one_iter()
  73. self.after_iter()
  74. def train_in_task(self):
  75. for task in self.train_loaders:
  76. self.before_task(task)
  77. self.train_in_iter(task)
  78. self.after_task()
  79. def before_task(self, train_loader):
  80. logger.info("init prefetcher, this might take one minute or less...")
  81. self.train_loader = train_loader
  82. self.prefetcher = DataPrefetcher(train_loader)
  83. self.learner = self.model.clone()
  84. def after_task(self):
  85. pass
  86. def adapt(self, inps, targets):
  87. # adapt_inps =inps[:1, ...]
  88. # targets_inps =targets[:1, ...]
  89. # print(adapt_inps.shape)
  90. # print(targets_inps.shape)
  91. outputs = self.learner(inps, targets)
  92. loss = outputs["total_loss"]
  93. self.learner.adapt(loss)
  94. def train_one_iter(self):
  95. iter_start_time = time.time()
  96. inps, targets = self.prefetcher.next()
  97. inps = inps.to(self.data_type)
  98. targets = targets.to(self.data_type)
  99. targets.requires_grad = False
  100. data_end_time = time.time()
  101. with torch.cuda.amp.autocast(enabled=self.amp_training):
  102. if self.iter % self.adaptation_period == 0:
  103. self.adapt(inps, targets)
  104. outputs = self.learner(inps, targets)
  105. loss = outputs["total_loss"]
  106. for p in self.exp.all_parameters:
  107. if p.grad is not None:
  108. p.grad.data.mul_(1.0 / self.args.batch_size)
  109. self.optimizer.zero_grad()
  110. self.scaler.scale(loss).backward()
  111. self.scaler.step(self.optimizer)
  112. self.scaler.update()
  113. if self.use_model_ema:
  114. self.ema_model.update(self.model)
  115. lr = self.lr_scheduler.update_lr(self.progress_in_iter + 1)
  116. for param_group in self.optimizer.param_groups:
  117. param_group["lr"] = lr
  118. iter_end_time = time.time()
  119. self.meter.update(
  120. iter_time=iter_end_time - iter_start_time,
  121. data_time=data_end_time - iter_start_time,
  122. lr=lr,
  123. **outputs,
  124. )
  125. def before_train(self):
  126. logger.info("args: {}".format(self.args))
  127. # logger.info("exp value:\n{}".format(self.exp))
  128. # model related init
  129. torch.cuda.set_device(self.local_rank)
  130. model = self.exp.get_model()
  131. logger.info(
  132. "Model Summary: {}".format(get_model_info(model, self.exp.test_size))
  133. )
  134. # exit()
  135. model.to(self.device)
  136. # from torchsummary import summary
  137. # summary(model, input_size=(3, 300, 300), device='cuda')
  138. # value of epoch will be set in `resume_train`
  139. model = self.resume_train(model)
  140. self.model = l2l.algorithms.MAML(model, lr=self.exp.inner_lr, first_order=self.exp.first_order)
  141. # solver related init
  142. self.optimizer = self.exp.get_optimizer(self.args.batch_size)
  143. # data related init
  144. self.no_aug = self.start_epoch >= self.max_epoch - self.exp.no_aug_epochs
  145. print('Getting data loaders')
  146. self.train_loaders = self.exp.get_data_loaders(
  147. batch_size=self.args.batch_size,
  148. is_distributed=self.is_distributed,
  149. no_aug=self.no_aug,
  150. )
  151. # max_iter means iters per epoch
  152. self.max_iter = 0
  153. for train_loader in self.train_loaders:
  154. self.max_iter += len(train_loader)
  155. self.lr_scheduler = self.exp.get_lr_scheduler(
  156. self.exp.basic_lr_per_img * self.args.batch_size, self.max_iter
  157. )
  158. if self.args.occupy:
  159. occupy_mem(self.local_rank)
  160. if self.is_distributed:
  161. self.model = DDP(self.model, device_ids=[self.local_rank], broadcast_buffers=False)
  162. if self.use_model_ema:
  163. self.ema_model = ModelEMA(self.model, 0.9998)
  164. self.ema_model.updates = self.max_iter * self.start_epoch
  165. # self.model = model
  166. self.model.train()
  167. self.evaluator = self.exp.get_evaluator(
  168. batch_size=self.args.batch_size, is_distributed=self.is_distributed
  169. )
  170. # Tensorboard logger
  171. if self.rank == 0:
  172. self.tblogger = SummaryWriter(self.file_name)
  173. logger.info("Training start...")
  174. # logger.info("\n{}".format(model))
  175. def after_train(self):
  176. logger.info(
  177. "Training of experiment is done and the best AP is {:.2f}".format(
  178. self.best_ap * 100
  179. )
  180. )
  181. def before_epoch(self):
  182. logger.info("---> start train epoch{}".format(self.epoch + 1))
  183. if self.epoch + 1 == self.max_epoch - self.exp.no_aug_epochs or self.no_aug:
  184. logger.info("--->No mosaic aug now!")
  185. for train_loader in self.train_loaders:
  186. train_loader.close_mosaic()
  187. logger.info("--->Add additional L1 loss now!")
  188. if self.is_distributed:
  189. self.model.module.head.use_l1 = True
  190. else:
  191. self.model.head.use_l1 = True
  192. self.exp.eval_interval = 1
  193. if not self.no_aug:
  194. self.save_ckpt(ckpt_name="last_mosaic_epoch")
  195. def after_epoch(self):
  196. if self.use_model_ema:
  197. self.ema_model.update_attr(self.model)
  198. self.save_ckpt(ckpt_name="latest")
  199. if (self.epoch + 1) % self.exp.eval_interval == 0:
  200. all_reduce_norm(self.model)
  201. self.evaluate_and_save_model()
  202. def before_iter(self):
  203. pass
  204. def after_iter(self):
  205. """
  206. `after_iter` contains two parts of logic:
  207. * log information
  208. * reset setting of resize
  209. """
  210. # log needed information
  211. if (self.iter + 1) % self.exp.print_interval == 0:
  212. # TODO check ETA logic
  213. left_iters = self.max_iter * self.max_epoch - (self.progress_in_iter + 1)
  214. eta_seconds = self.meter["iter_time"].global_avg * left_iters
  215. eta_str = "ETA: {}".format(datetime.timedelta(seconds=int(eta_seconds)))
  216. progress_str = "epoch: {}/{}, iter: {}/{}".format(
  217. self.epoch + 1, self.max_epoch, self.iter + 1, self.max_iter
  218. )
  219. loss_meter = self.meter.get_filtered_meter("loss")
  220. loss_str = ", ".join(
  221. ["{}: {:.3f}".format(k, v.latest) for k, v in loss_meter.items()]
  222. )
  223. time_meter = self.meter.get_filtered_meter("time")
  224. time_str = ", ".join(
  225. ["{}: {:.3f}s".format(k, v.avg) for k, v in time_meter.items()]
  226. )
  227. logger.info(
  228. "{}, mem: {:.0f}Mb, {}, {}, lr: {:.3e}".format(
  229. progress_str,
  230. gpu_mem_usage(),
  231. time_str,
  232. loss_str,
  233. self.meter["lr"].latest,
  234. )
  235. + (", size: {:d}, {}".format(self.input_size[0], eta_str))
  236. )
  237. self.meter.clear_meters()
  238. # random resizing
  239. if self.exp.random_size is not None and (self.progress_in_iter + 1) % 10 == 0:
  240. self.input_size = self.exp.random_resize(
  241. self.train_loader, self.epoch, self.rank, self.is_distributed
  242. )
  243. @property
  244. def progress_in_iter(self):
  245. return self.epoch * self.max_iter + self.iter
  246. def resume_train(self, model):
  247. if self.args.resume:
  248. logger.info("resume training")
  249. if self.args.ckpt is None:
  250. ckpt_file = os.path.join(self.file_name, "latest" + "_ckpt.pth.tar")
  251. else:
  252. ckpt_file = self.args.ckpt
  253. ckpt = torch.load(ckpt_file, map_location=self.device)
  254. # resume the model/optimizer state dict
  255. model.load_state_dict(ckpt["model"])
  256. self.optimizer.load_state_dict(ckpt["optimizer"])
  257. start_epoch = (
  258. self.args.start_epoch - 1
  259. if self.args.start_epoch is not None
  260. else ckpt["start_epoch"]
  261. )
  262. self.start_epoch = start_epoch
  263. logger.info(
  264. "loaded checkpoint '{}' (epoch {})".format(
  265. self.args.resume, self.start_epoch
  266. )
  267. ) # noqa
  268. else:
  269. if self.args.ckpt is not None:
  270. logger.info("loading checkpoint for fine tuning")
  271. ckpt_file = self.args.ckpt
  272. ckpt = torch.load(ckpt_file, map_location=self.device)["model"]
  273. model = load_ckpt(model, ckpt)
  274. self.start_epoch = 0
  275. return model
  276. def evaluate_and_save_model(self):
  277. evalmodel = self.ema_model.ema if self.use_model_ema else self.model
  278. ap50_95, ap50, summary = self.exp.eval(
  279. evalmodel, self.evaluator, self.is_distributed
  280. )
  281. self.model.train()
  282. if self.rank == 0:
  283. self.tblogger.add_scalar("val/COCOAP50", ap50, self.epoch + 1)
  284. self.tblogger.add_scalar("val/COCOAP50_95", ap50_95, self.epoch + 1)
  285. logger.info("\n" + summary)
  286. synchronize()
  287. # self.best_ap = max(self.best_ap, ap50_95)
  288. self.save_ckpt("last_epoch", ap50 > self.best_ap)
  289. self.best_ap = max(self.best_ap, ap50)
  290. def save_ckpt(self, ckpt_name, update_best_ckpt=False):
  291. if self.rank == 0:
  292. save_model = self.ema_model.ema if self.use_model_ema else self.model
  293. logger.info("Save weights to {}".format(self.file_name))
  294. ckpt_state = {
  295. "start_epoch": self.epoch + 1,
  296. "model": save_model.state_dict(),
  297. "optimizer": self.optimizer.state_dict(),
  298. }
  299. save_checkpoint(
  300. ckpt_state,
  301. update_best_ckpt,
  302. self.file_name,
  303. ckpt_name,
  304. )