| import copy as cp |
| import glob |
| from datetime import timedelta |
| from pathlib import Path |
| from omegaconf import OmegaConf |
| from omegaconf import open_dict |
| from tqdm import tqdm |
| import numpy as np |
|
|
| from accelerate import Accelerator, DistributedDataParallelKwargs |
| from accelerate.logging import get_logger |
| from accelerate.utils import set_seed, InitProcessGroupKwargs |
| from fvcore.common.registry import Registry |
| import torch |
| import wandb |
|
|
| import common.io_utils as iu |
| from common.io_utils import make_dir |
| import common.misc as misc |
| from data.build import build_dataloader |
| from evaluator.build import build_eval |
| from model.build import build_model |
| from optim.build import build_optim |
| from safetensors.torch import load_file |
|
|
| TRAINER_REGISTRY = Registry("Trainer") |
|
|
|
|
| def _global_l2(tensors): |
| """Compute sqrt(Σ‖t‖₂²) over an iterable of tensors.""" |
| total_sq = torch.tensor(0.0) |
| for t in tensors: |
| total_sq += t.float().pow(2).sum() |
| return total_sq.sqrt() |
|
|
|
|
| class Tracker(): |
| def __init__(self, cfg): |
| self.reset(cfg) |
|
|
| def step(self): |
| self.epoch += 1 |
|
|
| def reset(self, cfg): |
| self.exp_name = f"{cfg.exp_dir.parent.name.replace(f'{cfg.name}', '').lstrip('_')}/{cfg.exp_dir.name}" |
| self.epoch = 0 |
| self.best_result = -np.inf |
|
|
| def state_dict(self): |
| return {k: v for k, v in self.__dict__.items() if not k.startswith('__')} |
| |
| def load_state_dict(self, state_dict): |
| self.__dict__.update(state_dict) |
|
|
| @TRAINER_REGISTRY.register() |
| class BaseTrainer(): |
| def __init__(self, cfg): |
| set_seed(cfg.rng_seed) |
| self.debug = cfg.debug.get("flag", False) |
| self.hard_debug = cfg.debug.get("hard_debug", False) |
| self.epochs_per_eval = cfg.solver.get("epochs_per_eval", None) |
| self.epochs_per_save = cfg.solver.get("epochs_per_save", None) |
| self.global_step = 0 |
| |
| |
| self.exp_tracker = Tracker(cfg) |
| wandb_args = {"entity": cfg.logger.entity, "id": cfg.logger.run_id, "resume": cfg.resume} |
| if not cfg.logger.get('autoname'): |
| wandb_args["name"] = self.exp_tracker.exp_name |
| |
| self.logger = get_logger(__name__) |
| self.mode = cfg.mode |
|
|
| ddp_kwargs = DistributedDataParallelKwargs(find_unused_parameters=True) |
| init_kwargs = InitProcessGroupKwargs(timeout=timedelta(seconds=5400)) |
| kwargs = ([ddp_kwargs] if cfg.num_gpu > 1 else []) + [init_kwargs] |
|
|
| gradient_accumulation_steps = cfg.solver.get("gradient_accumulation_steps", 1) |
| |
| self.accelerator = Accelerator( |
| gradient_accumulation_steps=gradient_accumulation_steps, |
| log_with=cfg.logger.name, |
| kwargs_handlers=kwargs |
| ) |
| |
| if not self.hard_debug: |
| self.accelerator.init_trackers( |
| project_name=cfg.name if not self.debug else "Debug", |
| config=OmegaConf.to_container(cfg, resolve=True, throw_on_missing=True) if not cfg.resume else None, |
| init_kwargs={"wandb": wandb_args} |
| ) |
| |
| print(OmegaConf.to_yaml(cfg)) |
|
|
| if cfg.model.name == 'Query3D': |
| |
| |
| if 'mv' in cfg.model.memories or 'sem' in cfg.model.memories: |
| cfg.data.load_multiview_info = True |
| if 'voxel' in cfg.model.memories or 'sem' in cfg.model.memories: |
| cfg.data.load_mask3d_voxel = True |
| txt_model2tokenizer = {'BERTLanguageEncoder': 'bert-base-uncased', 'CLIPLanguageEncoder': 'openai/clip-vit-large-patch14'} |
| cfg.data_wrapper.tokenizer = txt_model2tokenizer[cfg.model.txt_encoder.name] |
| |
| if self.mode in ["warmup", "pretrain"]: |
| keys = [self.mode] |
| else: |
| keys = ["train", "val", "test"] |
| |
| self.data_loaders = {key : build_dataloader(cfg, split=key) for key in keys} |
| self.model = build_model(cfg) |
| |
| if self.mode == 'warmup': |
| self.epochs = cfg.solver.warmup_epochs |
| else: |
| self.epochs = cfg.solver.epochs |
| |
| if self.mode == "test": |
| total_steps = 1 |
| else: |
| total_steps = (len(self.data_loaders[self.mode]) * self.epochs) // gradient_accumulation_steps |
| self.loss, self.optimizer, self.scheduler = build_optim(cfg, self.model.get_opt_params(), |
| total_steps= total_steps, accelerator = self.accelerator) |
| |
| if misc.rgetattr(cfg, "eval.pass_kwargs", False): |
| kwargs = {"dataloaders": self.data_loaders} |
| else: |
| kwargs = {} |
| self.evaluator = build_eval(cfg, self.accelerator, **kwargs) |
|
|
| |
| |
| self.total_steps = 1 if self.mode == "test" else len(self.data_loaders[self.mode]) * self.epochs |
| self.grad_norm = cfg.solver.get("grad_norm") |
|
|
| ema = [0.996, 1.0] |
| ipe_scale = 1.0 |
| self.momentum_scheduler = (ema[0] + i*(ema[1]-ema[0])/(self.total_steps*self.epochs*ipe_scale) |
| for i in range(int(self.total_steps*self.epochs*ipe_scale)+1)) |
| |
| |
| if cfg.get('pretrain_ckpt_path'): |
| self.pretrain_ckpt_path = Path(cfg.pretrain_ckpt_path) |
| self.load_pretrain() |
| if hasattr(self.model, "pm_encoder"): |
| self.model.pm_encoder.load_state_dict(self.model.pm_encoder.state_dict()) |
|
|
| |
| self.model, self.loss, self.optimizer, self.scheduler = self.accelerator.prepare(self.model, self.loss, self.optimizer, self.scheduler) |
| for name, loader in self.data_loaders.items(): |
| if isinstance(loader, list): |
| loader = self.accelerator.prepare(*loader) |
| else: |
| loader = self.accelerator.prepare(loader) |
| self.data_loaders[name] = loader |
| self.accelerator.register_for_checkpointing(self.exp_tracker) |
|
|
| |
| self.ckpt_path = Path(cfg.ckpt_path) if cfg.get("ckpt_path") else Path(cfg.exp_dir) / "ckpt" / "best.pth" |
| if cfg.resume: |
| self.resume() |
|
|
| def forward(self, data_dict): |
| return self.model(data_dict) |
|
|
| def update_ema(self): |
| |
| with torch.no_grad(): |
| m = next(self.momentum_scheduler) |
| |
| model_context = self.model.module.context_model if hasattr(self.model, 'module') else self.model.context_model |
| model_target = self.model.module.target_model if hasattr(self.model, 'module') else self.model.target_model |
|
|
| for param_q, param_k in zip(model_context.parameters(), model_target.parameters()): |
| param_k.data.mul_(m).add_((1. - m) * param_q.detach().data) |
| |
| def backward(self, loss): |
| |
| self.accelerator.backward(loss) |
|
|
| total_norm = torch.norm(torch.stack([ |
| torch.norm(p.grad.detach()) for p in self.model.parameters() if p.grad is not None |
| ])) |
| print(f"grad_norm={total_norm.item():.2f}") |
| |
| |
| if self.grad_norm is not None and self.accelerator.sync_gradients: |
| self.accelerator.clip_grad_norm_(self.model.parameters(), self.grad_norm) |
|
|
| |
| if self.accelerator.sync_gradients: |
| self.optimizer.step() |
| self.optimizer.zero_grad() |
| self.scheduler.step() |
|
|
| def log(self, results, mode="train"): |
| if not self.hard_debug: |
| log_dict = {} |
| for key, val in results.items(): |
| if isinstance(val, torch.Tensor): |
| val = val.item() |
| log_dict[f"{mode}/{key}"] = val |
| if mode == "train": |
| lrs = self.scheduler.get_lr() |
| for i, lr in enumerate(lrs): |
| log_dict[f"{mode}/lr/group_{i}"] = lr |
| self.accelerator.log(log_dict, step=self.global_step) |
|
|
| def save(self, name): |
| make_dir(self.ckpt_path.parent) |
| self.save_func(str(self.ckpt_path.parent / name)) |
|
|
| def resume(self): |
| if self.ckpt_path.exists(): |
| print(f"Resuming from {str(self.ckpt_path)}") |
| |
| self.accelerator.load_state(str(self.ckpt_path)) |
| |
| print(f"Successfully resumed from {self.ckpt_path}") |
| else: |
| self.logger.info("training from scratch") |
| |
| def load_pretrain(self): |
| print(f"📂 Loading pretrained weights from: {str(self.pretrain_ckpt_path)}") |
| model_weight_path_pattern = str(self.pretrain_ckpt_path / "model*.safetensors") |
| model_weight_paths = glob.glob(model_weight_path_pattern) |
|
|
| if len(model_weight_paths) == 0: |
| raise FileNotFoundError(f"❌ Cannot find any .safetensors file in {str(self.pretrain_ckpt_path)}") |
|
|
| |
| weights = {} |
| for model_weight_path in model_weight_paths: |
| weights.update(load_file(model_weight_path, device="cpu")) |
|
|
| |
| result = self.model.load_state_dict(weights, strict=False) |
| |
| model_keys = set(self.model.state_dict().keys()) |
| loaded_keys = model_keys.intersection(weights.keys()) |
| missing_keys = result.missing_keys |
| unexpected_keys = result.unexpected_keys |
| print(missing_keys) |
| print(f"✅ Loaded keys: {len(loaded_keys)} / {len(model_keys)}") |
| print(f"❌ Missing keys: {len(missing_keys)}") |
| print(f"⚠️ Unexpected keys: {len(unexpected_keys)}") |
| |
| def save_func(self, path): |
| self.accelerator.save_state(path) |
| |
| def build_trainer(cfg): |
| return TRAINER_REGISTRY.get(cfg.trainer)(cfg) |